Files
secondo/bin/Scripts/SimilarityClusteringB.sec

249 lines
5.9 KiB
Plaintext
Raw Normal View History

2026-01-23 17:03:45 +08:00
# Similarity Clustering as in paper + balancing slots over workers.
# Database must be open.
# Distributed array T with attribute Pos of type point must be present.
# Workers relation must be present.
# Variable myPort must be set below to an exclusively used port.
# Parameter k below may be adapted (default 50)
# sample size in step 1 may be adapted
restore Workers from WorkersNewton;
let S = 'S' ffeed5 consume;
# 4:60 min
let T = S feed ddistribute3["T", 160, TRUE, Workers];
# 1:37 min
# prepare cost measurements
let ControlWorkers = createintdarray("ControlWorkers", Workers, Workers count)
@%Scripts/DistCost.sec
let myPort = ... ;
# Step 1
let sizeT = size(T);
query share("sizeT", TRUE, Workers)
let SS = T dmap["", . feed some[10000 div sizeT]] dsummarize consume
# Step 2
let k = 50;
@&Scripts/SimilarityPartitioning.sec;
let n = PC count;
let MinPts = 10;
let Eps = 100.0;
let wgs84 = create_geoid("WGS1984");
# Step 3
query share("PC", TRUE, Workers);
query share("MinPts", TRUE, Workers);
query share("Eps", TRUE, Workers);
query share("wgs84", TRUE, Workers);
query share("n", TRUE, Workers);
# Step 4
query memclear();
query T dcommand['query meminit(3600)'] consume;
query T dlet["PCm", 'PC feed mconsume'] consume;
query T dlet["PCm_Pos_mtree", 'PCm mcreatemtree[Pos, wgs84]'] consume
let Va = T
dmap["", . feed
loopjoin[fun(t: TUPLE) PCm_Pos_mtree PCm mdistScan[attr(t, Pos)] head[1]
projectextend[N; Dist: distance(attr(t, Pos), .Pos, wgs84)]]
loopjoin[fun(u: TUPLE) PCm_Pos_mtree PCm mdistRange[attr(u, Pos),
attr(u, Dist) + (2 * Eps)] projectextend[; N2: .N]]
]
partition["Va", .N2, n]
let Vb = Va
collect2["Vb", myPort]
# Step 4b Load Balancing
let wc = Workers count;
let reserve = ((wc - 1) div 20) + 1;
let W = intstream(0, wc - 1) namedtransformstream[Worker] extend[Load: 0.0] consume
let Sizes = Vb dmap["", . feed count] dsummarize namedtransformstream[Size] addcounter[Slot, 0] consume
let Slots = Sizes feed replaceAttr[Size: .Size * 1.0] sortby[Size desc] consume
let TargetSize = (Slots feed sum[Size]) / wc;
query memclear();
let PQ = W feed head[wc - reserve] mcreatepqueue[Load];
let Assignment = PQ mfeedpq
Slots feed obojoin
extend[Ok: PQ minserttuplepqprojectU[., .Load + .Size, Load; Worker, Load]]
consume
delete PQ;
let PQ = W feed tail[reserve] mcreatepqueue[Load];
query PQ mfeedpq
Assignment feed addid extend[LoadAfter: .Load + .Size] sortby[LoadAfter desc]
project[TID]
Assignment deletebyid2[TID]
project[Size, Slot]
obojoin
extend[Ok: PQ minserttuplepqprojectU[., .Load + .Size, Load; Worker, Load]]
Assignment insert
cancel[(.Load + (2 * .Size)) > (TargetSize * 1.03)]
count
let AssignmentV = Assignment feed sortby[Slot] project[Worker] transformstream collect_vector
let V = Va collectC["V", myPort, AssignmentV]
# Step 5
update LastCommand := distCostReset(ControlWorkers)
let X = V
dmap["X", $1 feed extend[Pos2: gk(.Pos)] dbscanM[Pos2, CID0, Eps, MinPts]
extend[CID: (.CID0 * n) + $2] consume
]
let Cost1 = distCostSave(ControlWorkers);
update LastCommand := distCostReset(ControlWorkers)
# Step 6
query T dcommand['query memclear()'] filter[.Ok] count;
let Wm = X dmap["Wm", . feed filter[.N = .N2] mconsume];
let Wm_Pos_mtree = Wm dmap["Wm_Pos_mtree", . mcreatemtree[Pos, wgs84]];
let Neighbors = X Wm_Pos_mtree Wm
dmap3["Neighbors", $1 feed filter[.N # .N2]
loopsel[fun(t: TUPLE) $2 $3 mdistRange[attr(t, Pos), Eps]
projectextend[; P: .Osm_id, PosP: .Pos, CID0: .CID0, CIDp: .CID,
IsCoreP: .IsCore, Np: .N, Q: attr(t, Osm_id), QPos: attr(t, Pos)]]
, myPort
]
let Cost2 = distCostSave(ControlWorkers);
update LastCommand := distCostReset(ControlWorkers)
query T dcommand['query memclear()'] filter[.Ok] count;
let NeighborsByP = Neighbors partition["", hashvalue(.P, 999997), 0]
collect2["NeighborsByP", myPort];
let NeighborsByQ = Neighbors partition["", hashvalue(.Q, 999997), 0]
collect2["NeighborsByQ", myPort];
let Cost3 = distCostSave(ControlWorkers);
update LastCommand := distCostReset(ControlWorkers)
# Step 7
let Merge = NeighborsByQ NeighborsByP
dmap2["Merge", . feed {n1} .. feed {n2} itHashJoin[Q_n1, P_n2]
filter[.IsCoreP_n1 and .IsCoreP_n2]
project[CIDp_n1, CIDp_n2] sort rdup, myPort
]
let Cost4 = distCostSave(ControlWorkers);
update LastCommand := distCostReset(ControlWorkers)
let Assignments = NeighborsByQ NeighborsByP
dmap2["", . feed {n1} .. feed {n2} itHashJoin[Q_n1, P_n2], myPort]
dmap["",
. feed filter[.IsCoreP_n1 and not(.IsCoreP_n2)]
projectextend[; P: .P_n2, N: .Np_n2, CID: .CIDp_n1]
. feed filter[.IsCoreP_n2 and not(.IsCoreP_n1)]
projectextend[; P: .P_n1, N: .Np_n1, CID: .CIDp_n2]
concat sort krdup[P]
]
partition["", .N, 0]
collect2["Assignments", myPort]
let Cost5 = distCostSave(ControlWorkers);
update LastCommand := distCostReset(ControlWorkers)
# Step 8
let MergeM = Merge dsummarize sort rdup createmgraph2[CIDp_n1, CIDp_n2, 1.0];
let MaxCN = X dmap["", . feed max[CID] feed transformstream] dsummarize max[Elem];
# Step 9
let Renumber = MergeM mg2connectedcomponents projectextend[; CID: .CIDp_n1,
CIDnew: .CompNo + MaxCN] sort rdup consume
# Step 10
query share("Renumber", TRUE, Workers);
# Step 11
update LastCommand := distCostReset(ControlWorkers)
query X Assignments
dmap2["", $1 feed addid filter[.N = .N2] $2 feed sort krdup[P] {a}
itHashJoin[Osm_id, P_a] $1 updatedirect2[TID; CID: .CID_a] count, myPort
]
getValue tie[. + ..]
let Cost6 = distCostSave(ControlWorkers);
update LastCommand := distCostReset(ControlWorkers)
query X
dmap["", $1 feed addid filter[.N = .N2] Renumber feed sort krdup[CID] {a}
itHashJoin[CID, CID_a] $1 updatedirect2[TID; CID: .CIDnew_a] count
]
getValue tie[. + ..]
let Cost7 = distCostSave(ControlWorkers);
update LastCommand := distCostReset(ControlWorkers)
let Commands = SEC2COMMANDS feed consume