Merge branch 'master' of https://github.com/mapnik/mapnik
This commit is contained in:
commit
afae890dcb
9 changed files with 109 additions and 14 deletions
|
@ -94,6 +94,8 @@ Released Aug 23, 2012
|
|||
|
||||
- Added support for justify-alignment=auto. This is the new default. (#1125)
|
||||
|
||||
- Added support for grouped rendering using the `group-by` layer option: https://github.com/mapnik/mapnik/wiki/Grouped-rendering
|
||||
|
||||
|
||||
## Mapnik 2.0.2
|
||||
|
||||
|
|
|
@ -143,7 +143,7 @@ void csv_datasource::bind() const
|
|||
}
|
||||
|
||||
template <typename T>
|
||||
void csv_datasource::parse_csv(T& stream,
|
||||
void csv_datasource::parse_csv(T & stream,
|
||||
std::string const& escape,
|
||||
std::string const& separator,
|
||||
std::string const& quote) const
|
||||
|
@ -171,6 +171,7 @@ void csv_datasource::parse_csv(T& stream,
|
|||
|
||||
// autodetect newlines
|
||||
char newline = '\n';
|
||||
bool has_newline = false;
|
||||
int newline_count = 0;
|
||||
int carriage_count = 0;
|
||||
for (unsigned idx = 0; idx < file_length_; idx++)
|
||||
|
@ -179,10 +180,12 @@ void csv_datasource::parse_csv(T& stream,
|
|||
if (c == '\n')
|
||||
{
|
||||
++newline_count;
|
||||
has_newline = true;
|
||||
}
|
||||
else if (c == '\r')
|
||||
{
|
||||
++carriage_count;
|
||||
has_newline = true;
|
||||
}
|
||||
// read at least 2000 bytes before testing
|
||||
if (idx == file_length_-1 || idx > 4000)
|
||||
|
@ -422,7 +425,7 @@ void csv_datasource::parse_csv(T& stream,
|
|||
throw mapnik::datasource_exception(s.str());
|
||||
}
|
||||
|
||||
int feature_count(1);
|
||||
int feature_count(0);
|
||||
bool extent_initialized = false;
|
||||
std::size_t num_headers = headers_.size();
|
||||
|
||||
|
@ -435,12 +438,23 @@ void csv_datasource::parse_csv(T& stream,
|
|||
mapnik::wkt_parser parse_wkt;
|
||||
mapnik::json::geometry_parser<std::string::const_iterator> parse_json;
|
||||
|
||||
while (std::getline(stream,csv_line,newline))
|
||||
// handle rare case of a single line of data and user-provided headers
|
||||
// where a lack of a newline will mean that std::getline returns false
|
||||
bool is_first_row = false;
|
||||
if (!has_newline)
|
||||
{
|
||||
stream >> csv_line;
|
||||
if (!csv_line.empty())
|
||||
{
|
||||
is_first_row = true;
|
||||
}
|
||||
}
|
||||
while (std::getline(stream,csv_line,newline) || is_first_row)
|
||||
{
|
||||
is_first_row = false;
|
||||
if ((row_limit_ > 0) && (line_number > row_limit_))
|
||||
{
|
||||
MAPNIK_LOG_DEBUG(csv) << "csv_datasource: row limit hit, exiting at feature: " << feature_count;
|
||||
|
||||
break;
|
||||
}
|
||||
|
||||
|
@ -495,7 +509,8 @@ void csv_datasource::parse_csv(T& stream,
|
|||
}
|
||||
}
|
||||
|
||||
mapnik::feature_ptr feature(mapnik::feature_factory::create(ctx_,feature_count));
|
||||
// NOTE: we use ++feature_count here because feature id's should start at 1;
|
||||
mapnik::feature_ptr feature(mapnik::feature_factory::create(ctx_,++feature_count));
|
||||
double x(0);
|
||||
double y(0);
|
||||
bool parsed_x = false;
|
||||
|
@ -754,7 +769,6 @@ void csv_datasource::parse_csv(T& stream,
|
|||
extent_.expand_to_include(feature->envelope());
|
||||
}
|
||||
features_.push_back(feature);
|
||||
++feature_count;
|
||||
null_geom = false;
|
||||
}
|
||||
else
|
||||
|
@ -782,7 +796,6 @@ void csv_datasource::parse_csv(T& stream,
|
|||
pt->move_to(x,y);
|
||||
feature->add_geometry(pt);
|
||||
features_.push_back(feature);
|
||||
++feature_count;
|
||||
null_geom = false;
|
||||
if (!extent_initialized)
|
||||
{
|
||||
|
@ -836,6 +849,9 @@ void csv_datasource::parse_csv(T& stream,
|
|||
else
|
||||
{
|
||||
MAPNIK_LOG_ERROR(csv) << s.str();
|
||||
// with no geometry we will never
|
||||
// add this feature so drop the count
|
||||
feature_count--;
|
||||
continue;
|
||||
}
|
||||
}
|
||||
|
|
|
@ -54,7 +54,7 @@ public:
|
|||
void bind() const;
|
||||
|
||||
template <typename T>
|
||||
void parse_csv(T& stream,
|
||||
void parse_csv(T & stream,
|
||||
std::string const& escape,
|
||||
std::string const& separator,
|
||||
std::string const& quote) const;
|
||||
|
|
|
@ -108,9 +108,7 @@ void agg_renderer<T>::process(line_symbolizer const& sym,
|
|||
{
|
||||
typedef agg::renderer_outline_aa<renderer_base> renderer_type;
|
||||
typedef agg::rasterizer_outline_aa<renderer_type> rasterizer_type;
|
||||
// need to reduce width by half to match standard rasterizer look
|
||||
double scaled = scale_factor_ * .5;
|
||||
agg::line_profile_aa profile(stroke_.get_width() * scaled, agg::gamma_power(stroke_.get_gamma()));
|
||||
agg::line_profile_aa profile(stroke_.get_width() * scale_factor_, agg::gamma_power(stroke_.get_gamma()));
|
||||
renderer_type ren(renb, profile);
|
||||
ren.color(agg::rgba8_pre(r, g, b, int(a*stroke_.get_opacity())));
|
||||
rasterizer_type ras(ren);
|
||||
|
@ -118,14 +116,12 @@ void agg_renderer<T>::process(line_symbolizer const& sym,
|
|||
|
||||
vertex_converter<box2d<double>, rasterizer_type, line_symbolizer,
|
||||
CoordTransform, proj_transform, agg::trans_affine, conv_types>
|
||||
converter(clipping_extent,ras,sym,t_,prj_trans,tr,scaled);
|
||||
converter(clipping_extent,ras,sym,t_,prj_trans,tr,scale_factor_);
|
||||
if (sym.clip()) converter.set<clip_line_tag>(); // optional clip (default: true)
|
||||
converter.set<transform_tag>(); // always transform
|
||||
if (fabs(sym.offset()) > 0.0) converter.set<offset_transform_tag>(); // parallel offset
|
||||
converter.set<affine_transform_tag>(); // optional affine transform
|
||||
if (sym.smooth() > 0.0) converter.set<smooth_tag>(); // optional smooth converter
|
||||
if (stroke_.has_dash()) converter.set<dash_tag>();
|
||||
converter.set<stroke_tag>(); //always stroke
|
||||
|
||||
BOOST_FOREACH( geometry_type & geom, feature.paths())
|
||||
{
|
||||
|
|
1
tests/data/csv/fails/needs_headers_one_line.csv
Normal file
1
tests/data/csv/fails/needs_headers_one_line.csv
Normal file
|
@ -0,0 +1 @@
|
|||
0,0,data_name
|
|
|
@ -0,0 +1 @@
|
|||
0,0,data_name
|
|
2
tests/data/csv/fails/needs_headers_two_lines.csv
Normal file
2
tests/data/csv/fails/needs_headers_two_lines.csv
Normal file
|
@ -0,0 +1,2 @@
|
|||
0,0,data_name
|
||||
0,0,data_name
|
|
4
tests/data/csv/warns/feature_id_counting.csv
Normal file
4
tests/data/csv/warns/feature_id_counting.csv
Normal file
|
@ -0,0 +1,4 @@
|
|||
x,y,id
|
||||
0,0,1
|
||||
bad,bad,2
|
||||
0,0,2
|
|
|
@ -418,6 +418,79 @@ if 'csv' in mapnik.DatasourceCache.instance().plugin_names():
|
|||
# this has invalid header # so throw
|
||||
ds = get_csv_ds('more_column_values_than_headers.csv')
|
||||
|
||||
def test_that_feature_id_only_incremented_for_valid_rows(**kwargs):
|
||||
ds = mapnik.Datasource(type='csv',
|
||||
file=os.path.join('../data/csv/warns','feature_id_counting.csv'),
|
||||
quiet=True)
|
||||
eq_(len(ds.fields()),3)
|
||||
eq_(ds.fields(),['x','y','id'])
|
||||
eq_(ds.field_types(),['int','int','int'])
|
||||
fs = ds.featureset()
|
||||
# first
|
||||
feat = fs.next()
|
||||
eq_(feat['x'],0)
|
||||
eq_(feat['y'],0)
|
||||
eq_(feat['id'],1)
|
||||
# second, should have skipped bogus one
|
||||
feat = fs.next()
|
||||
eq_(feat['x'],0)
|
||||
eq_(feat['y'],0)
|
||||
eq_(feat['id'],2)
|
||||
desc = ds.describe()
|
||||
eq_(desc['geometry_type'],mapnik.DataGeometryType.Point)
|
||||
eq_(len(ds.all_features()),2)
|
||||
|
||||
def test_dynamically_defining_headers1(**kwargs):
|
||||
ds = mapnik.Datasource(type='csv',
|
||||
file=os.path.join('../data/csv/fails','needs_headers_two_lines.csv'),
|
||||
quiet=True,
|
||||
headers='x,y,name')
|
||||
eq_(len(ds.fields()),3)
|
||||
eq_(ds.fields(),['x','y','name'])
|
||||
eq_(ds.field_types(),['int','int','str'])
|
||||
fs = ds.featureset()
|
||||
feat = fs.next()
|
||||
eq_(feat['x'],0)
|
||||
eq_(feat['y'],0)
|
||||
eq_(feat['name'],'data_name')
|
||||
desc = ds.describe()
|
||||
eq_(desc['geometry_type'],mapnik.DataGeometryType.Point)
|
||||
eq_(len(ds.all_features()),2)
|
||||
|
||||
def test_dynamically_defining_headers2(**kwargs):
|
||||
ds = mapnik.Datasource(type='csv',
|
||||
file=os.path.join('../data/csv/fails','needs_headers_one_line.csv'),
|
||||
quiet=True,
|
||||
headers='x,y,name')
|
||||
eq_(len(ds.fields()),3)
|
||||
eq_(ds.fields(),['x','y','name'])
|
||||
eq_(ds.field_types(),['int','int','str'])
|
||||
fs = ds.featureset()
|
||||
feat = fs.next()
|
||||
eq_(feat['x'],0)
|
||||
eq_(feat['y'],0)
|
||||
eq_(feat['name'],'data_name')
|
||||
desc = ds.describe()
|
||||
eq_(desc['geometry_type'],mapnik.DataGeometryType.Point)
|
||||
eq_(len(ds.all_features()),1)
|
||||
|
||||
def test_dynamically_defining_headers3(**kwargs):
|
||||
ds = mapnik.Datasource(type='csv',
|
||||
file=os.path.join('../data/csv/fails','needs_headers_one_line_no_newline.csv'),
|
||||
quiet=True,
|
||||
headers='x,y,name')
|
||||
eq_(len(ds.fields()),3)
|
||||
eq_(ds.fields(),['x','y','name'])
|
||||
eq_(ds.field_types(),['int','int','str'])
|
||||
fs = ds.featureset()
|
||||
feat = fs.next()
|
||||
eq_(feat['x'],0)
|
||||
eq_(feat['y'],0)
|
||||
eq_(feat['name'],'data_name')
|
||||
desc = ds.describe()
|
||||
eq_(desc['geometry_type'],mapnik.DataGeometryType.Point)
|
||||
eq_(len(ds.all_features()),1)
|
||||
|
||||
if __name__ == "__main__":
|
||||
setup()
|
||||
[eval(run)(visual=True) for run in dir() if 'test_' in run]
|
||||
|
|
Loading…
Reference in a new issue