forked from osm2pgsql-dev/osm2pgsql
-
Notifications
You must be signed in to change notification settings - Fork 0
/
Copy pathmiddle-ram.cpp
429 lines (355 loc) · 12.1 KB
/
middle-ram.cpp
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
/* Implements the mid-layer processing for osm2pgsql
* using several arrays in RAM. This is fastest if you
* have sufficient RAM+Swap.
*
* This layer stores data read in from the planet.osm file
* and is then read by the backend processing code to
* emit the final geometry-enabled output formats
*/
#include <stdio.h>
#include <unistd.h>
#include <stdlib.h>
#include <string.h>
#include <assert.h>
#include <libpq-fe.h>
#include <stdexcept>
#include "osmtypes.hpp"
#include "middle-ram.hpp"
#include "node-ram-cache.hpp"
#include "output-pgsql.hpp"
#include "options.hpp"
#include "util.hpp"
/* Store +-20,000km Mercator co-ordinates as fixed point 32bit number with maximum precision */
/* Scale is chosen such that 40,000 * SCALE < 2^32 */
#define FIXED_POINT
/* Object storage now uses 2 levels of storage arrays.
*
* - Low level storage of 2^16 (~65k) objects in an indexed array
* These are allocated dynamically when we need to first store data with
* an ID in this block
*
* - Fixed array of 2^(32 - 16) = 65k pointers to the dynamically allocated arrays.
*
* This allows memory usage to be efficient and scale dynamically without needing to
* hard code maximum IDs. We now support an ID range of -2^31 to +2^31.
* The negative IDs often occur in non-uploaded JOSM data or other data import scripts.
*
*/
#define BLOCK_SHIFT 10
#define PER_BLOCK (1 << BLOCK_SHIFT)
#define NUM_BLOCKS (1 << (32 - BLOCK_SHIFT))
static osmid_t id2block(osmid_t id)
{
/* + NUM_BLOCKS/2 allows for negative IDs */
return (id >> BLOCK_SHIFT) + NUM_BLOCKS/2;
}
static osmid_t id2offset(osmid_t id)
{
return id & (PER_BLOCK-1);
}
static int block2id(int block, int offset)
{
return ((block - NUM_BLOCKS/2) << BLOCK_SHIFT) + offset;
}
#define UNUSED __attribute__ ((unused))
int middle_ram_t::nodes_set(osmid_t id, double lat, double lon, struct keyval *tags) {
return cache->set(id, lat, lon, tags);
}
int middle_ram_t::ways_set(osmid_t id, osmid_t *nds, int nd_count, struct keyval *tags)
{
int pending = 0;
int block = id2block(id);
int offset = id2offset(id);
struct keyval *p;
if (!ways[block]) {
ways[block] = (struct ramWay *)calloc(PER_BLOCK, sizeof(struct ramWay));
if (!ways[block]) {
fprintf(stderr, "Error allocating ways\n");
util::exit_nicely();
}
way_blocks++;
}
if (ways[block][offset].ndids) {
free(ways[block][offset].ndids);
ways[block][offset].ndids = NULL;
}
/* Copy into length prefixed array */
ways[block][offset].ndids = (osmid_t *)malloc( (nd_count+1)*sizeof(osmid_t) );
memcpy( ways[block][offset].ndids+1, nds, nd_count*sizeof(osmid_t) );
ways[block][offset].ndids[0] = nd_count;
ways[block][offset].pending = pending;
if (!ways[block][offset].tags) {
p = (struct keyval *)malloc(sizeof(struct keyval));
if (p) {
initList(p);
ways[block][offset].tags = p;
} else {
fprintf(stderr, "%s malloc failed\n", __FUNCTION__);
util::exit_nicely();
}
} else
resetList(ways[block][offset].tags);
cloneList(ways[block][offset].tags, tags);
return 0;
}
int middle_ram_t::relations_set(osmid_t id, struct member *members, int member_count, struct keyval *tags)
{
struct keyval *p;
struct member *ptr;
int block = id2block(id);
int offset = id2offset(id);
if (!rels[block]) {
rels[block] = (struct ramRel *)calloc(PER_BLOCK, sizeof(struct ramRel));
if (!rels[block]) {
fprintf(stderr, "Error allocating rels\n");
util::exit_nicely();
}
}
if (!rels[block][offset].tags) {
p = (struct keyval *)malloc(sizeof(struct keyval));
if (p) {
initList(p);
rels[block][offset].tags = p;
} else {
fprintf(stderr, "%s malloc failed\n", __FUNCTION__);
util::exit_nicely();
}
} else
resetList(rels[block][offset].tags);
cloneList(rels[block][offset].tags, tags);
if (!rels[block][offset].members)
free( rels[block][offset].members );
ptr = (struct member *)malloc(sizeof(struct member) * member_count);
if (ptr) {
memcpy( ptr, members, sizeof(struct member) * member_count );
rels[block][offset].member_count = member_count;
rels[block][offset].members = ptr;
} else {
fprintf(stderr, "%s malloc failed\n", __FUNCTION__);
util::exit_nicely();
}
return 0;
}
int middle_ram_t::nodes_get_list(struct osmNode *nodes, const osmid_t *ndids, int nd_count) const
{
int i, count;
count = 0;
for( i=0; i<nd_count; i++ )
{
if (cache->get(&nodes[count], ndids[i]))
continue;
count++;
}
return count;
}
void middle_ram_t::iterate_relations(middle_t::cb_func &callback)
{
int block, offset;
// to maintain backwards compatibility, we need to set this flag
// which fakes the previous behaviour of having deleted all the
// ways.
simulate_ways_deleted = true;
fprintf(stderr, "\n");
for(block=NUM_BLOCKS-1; block>=0; block--) {
if (!rels[block])
continue;
for (offset=0; offset < PER_BLOCK; offset++) {
if (rels[block][offset].members) {
osmid_t id = block2id(block, offset);
rel_out_count++;
if (rel_out_count % 10 == 0)
fprintf(stderr, "\rWriting relation (%u)", rel_out_count);
callback(id, 0);
}
}
}
fprintf(stderr, "\rWriting relation (%u)\n", rel_out_count);
}
size_t middle_ram_t::pending_count() const {
//TODO: keep a running count of marked pending stuff
//so we dont have to iterate over the memory to know
return 42;
}
void middle_ram_t::iterate_ways(middle_t::pending_processor& pf)
{
int block, offset;
fprintf(stderr, "\n");
for(block=NUM_BLOCKS-1; block>=0; block--) {
if (!ways[block])
continue;
for (offset=0; offset < PER_BLOCK; offset++) {
if (ways[block][offset].ndids) {
way_out_count++;
if (way_out_count % 1000 == 0)
fprintf(stderr, "\rEnqueuing way (%uk)", way_out_count/1000);
if (ways[block][offset].pending) {
/* First element contains number of nodes */
if (ways[block][offset].ndids[0]) {
osmid_t id = block2id(block, offset);
pf.enqueue(id);
}
ways[block][offset].pending = 0;
}
}
}
}
fprintf(stderr, "\rEnqueuing way (%uk)\n", way_out_count/1000);
//let the threads process the ways
pf.process_ways();
//TODO: message to show the real progress of writing the ways
}
void middle_ram_t::release_relations()
{
int block, offset;
for(block=NUM_BLOCKS-1; block>=0; block--) {
if (!rels[block])
continue;
for (offset=0; offset < PER_BLOCK; offset++) {
if (rels[block][offset].members) {
free(rels[block][offset].members);
rels[block][offset].members = NULL;
resetList(rels[block][offset].tags);
free(rels[block][offset].tags);
rels[block][offset].tags=NULL;
}
}
free(rels[block]);
rels[block] = NULL;
}
}
void middle_ram_t::release_ways()
{
int i, j = 0;
for (i=0; i<NUM_BLOCKS; i++) {
if (ways[i]) {
for (j=0; j<PER_BLOCK; j++) {
if (ways[i][j].tags) {
resetList(ways[i][j].tags);
free(ways[i][j].tags);
}
if (ways[i][j].ndids)
free(ways[i][j].ndids);
}
free(ways[i]);
ways[i] = NULL;
}
}
}
/* Caller must free nodes_ptr and resetList(tags_ptr) */
int middle_ram_t::ways_get(osmid_t id, struct keyval *tags_ptr, struct osmNode **nodes_ptr, int *count_ptr) const
{
int block = id2block(id), offset = id2offset(id), ndCount = 0;
struct osmNode *nodes;
if (simulate_ways_deleted)
return 1;
if (!ways[block])
return 1;
if (ways[block][offset].ndids) {
/* First element contains number of nodes */
nodes = (struct osmNode *)malloc( sizeof(struct osmNode) * ways[block][offset].ndids[0]);
ndCount = nodes_get_list(nodes, ways[block][offset].ndids+1, ways[block][offset].ndids[0]);
if (ndCount) {
cloneList( tags_ptr, ways[block][offset].tags );
*nodes_ptr = nodes;
*count_ptr = ndCount;
return 0;
}
free(nodes);
}
return 1;
}
int middle_ram_t::ways_get_list(const osmid_t *ids, int way_count, osmid_t *way_ids, struct keyval *tag_ptr, struct osmNode **node_ptr, int *count_ptr) const {
int count = 0;
int i;
initList(&(tag_ptr[count]));
for (i = 0; i < way_count; i++) {
if (ways_get(ids[i], &(tag_ptr[count]), &(node_ptr[count]), &(count_ptr[count])) == 0) {
way_ids[count] = ids[i];
count++;
initList(&(tag_ptr[count]));
}
}
return count;
}
/* Caller must free members_ptr and resetList(tags_ptr).
* Note that the members in members_ptr are copied, but the roles
* within the members are not, and should not be freed.
*/
int middle_ram_t::relations_get(osmid_t id, struct member **members_ptr, int *member_count, struct keyval *tags_ptr) const
{
int block = id2block(id), offset = id2offset(id), ndCount = 0;
struct member *members;
if (!rels[block])
return 1;
if (rels[block][offset].members) {
const size_t member_bytes = sizeof(struct member) * rels[block][offset].member_count;
members = (struct member *)malloc(member_bytes);
memcpy(members, rels[block][offset].members, member_bytes);
cloneList( tags_ptr, rels[block][offset].tags );
*members_ptr = members;
*member_count = rels[block][offset].member_count;
return 0;
}
return 1;
}
void middle_ram_t::analyze(void)
{
/* No need */
}
void middle_ram_t::cleanup(void)
{
/* No need */
}
void middle_ram_t::end(void)
{
/* No need */
}
int middle_ram_t::start(const options_t *out_options_)
{
out_options = out_options_;
/* latlong has a range of +-180, mercator +-20000
The fixed poing scaling needs adjusting accordingly to
be stored accurately in an int */
cache.reset(new node_ram_cache(out_options->alloc_chunkwise, out_options->cache, out_options->scale));
fprintf( stderr, "Mid: Ram, scale=%d\n", out_options->scale );
return 0;
}
void middle_ram_t::stop(void)
{
int i, j;
cache.reset(NULL);
release_ways();
release_relations();
}
void middle_ram_t::commit(void) {
}
middle_ram_t::middle_ram_t():
ways(), rels(), way_blocks(0), way_out_count(0), rel_out_count(0), cache(),
simulate_ways_deleted(false)
{
ways.resize(NUM_BLOCKS); memset(&ways[0], 0, NUM_BLOCKS * sizeof ways[0]);
rels.resize(NUM_BLOCKS); memset(&rels[0], 0, NUM_BLOCKS * sizeof rels[0]);
}
middle_ram_t::~middle_ram_t() {
//instance.reset();
}
std::vector<osmid_t> middle_ram_t::relations_using_way(osmid_t way_id) const
{
// this function shouldn't be called - relations_using_way is only used in
// slim mode, and a middle_ram_t shouldn't be constructed if the slim mode
// option is set.
throw std::runtime_error("middle_ram_t::relations_using_way is unimlpemented, and "
"should not have been called. This is probably a bug, please "
"report it at https://github.com/openstreetmap/osm2pgsql/issues");
}
namespace {
void no_delete(const middle_ram_t * middle) {
// boost::shared_ptr thinks we are going to delete
// the middle object, but we are not. Heh heh heh.
// So yeah, this is a hack...
}
}
boost::shared_ptr<const middle_query_t> middle_ram_t::get_instance() const {
//shallow copy here because readonly access is thread safe
return boost::shared_ptr<const middle_query_t>(this, no_delete);
}