private void countIgnoredFields() { for (FieldMetaData field : fields) { if (field.ignore()) { ignoredFieldCount++; } else { unignoredFieldCount++; } } }
/** * Get the unignored field count for the specified record. * * @param recordMetaData The record metadata. * @return The unignored field count. */ public int getUnignoredFieldCount(RecordMetaData recordMetaData) { if (factory.isMultiTypeRecordSet()) { // Need to account for the leading identifier field on each // record... return recordMetaData.getUnignoredFieldCount() + 1; } else { return recordMetaData.getUnignoredFieldCount(); } }
private void addFieldBindings(Bean bean) { for (FieldMetaData fieldMetaData : vfRecordMetaData.getRecordMetaData().getFields()) { if (!fieldMetaData.ignore()) { bean.bindTo(fieldMetaData.getName(), recordElementName + "/" + fieldMetaData.getName()); } } }
protected void validateHeader(List<String> headers) throws IOException { if (factory.isMultiTypeRecordSet()) { throw new IOException("Cannot validate the 'header' field of a Multi-Type Record Set. Reader fields definition defines multiple record definitions."); } RecordMetaData recordMetaData = factory.getRecordMetaData(); if (headers == null) { throw new IOException("Null header."); } if (validateHeader(headers, recordMetaData.getFields())) { return; } throw new IOException("Invalid header."); }
@Override protected void validateHeader(List<String> headers) { // For backward compatibility with pre v1.5.... try { super.validateHeader(headers); } catch (IOException e) { throw new CSVHeaderValidationException(getFactory().getRecordMetaData().getFieldNames(), headers); } } }
public CSVMapBinder(String fields, Class recordType, String keyField) { AssertArgument.isNotNullAndNotEmpty(fields, "fields"); AssertArgument.isNotNull(recordType, "recordType"); AssertArgument.isNotNullAndNotEmpty(keyField, "keyField"); smooks = new Smooks(); smooks.setFilterSettings(FilterSettings.DEFAULT_SAX); smooks.setReaderConfig(new CSVRecordParserConfigurator(fields) .setBinding(new Binding(beanId, recordType, BindingType.MAP).setKeyField(keyField))); }
/** * public constructor. * * @param name Record name. * @param fields Record fields metadata. * @param wildCardRecord Wildcard record. Accept any fields and generate the * field names based on index. */ public RecordMetaData(String name, List<FieldMetaData> fields, boolean wildCardRecord) { AssertArgument.isNotNullAndNotEmpty(name, "name"); AssertArgument.isNotNull(fields, "fields"); this.name = name.trim(); this.fields = fields; this.wildCardRecord = wildCardRecord; countIgnoredFields(); gatherFieldNames(); }
public FieldMetaData(String name) { AssertArgument.isNotNullAndNotEmpty(name, "name"); this.name = name; ignore = name.startsWith(IGNORE_FIELD); if (ignore) { ignoreCount = parseIgnoreFieldDirective(name); } }
public CSVListBinder(String fields, Class recordType) { AssertArgument.isNotNullAndNotEmpty(fields, "fields"); AssertArgument.isNotNull(recordType, "recordType"); smooks = new Smooks(); smooks.setFilterSettings(FilterSettings.DEFAULT_SAX); smooks.setReaderConfig(new CSVRecordParserConfigurator(fields) .setBinding(new Binding(beanId, recordType, BindingType.LIST))); }
protected void validateHeader(List<String> headers) throws IOException { if (factory.isMultiTypeRecordSet()) { throw new IOException("Cannot validate the 'header' field of a Multi-Type Record Set. Reader fields definition defines multiple record definitions."); } RecordMetaData recordMetaData = factory.getRecordMetaData(); if (headers == null) { throw new IOException("Null header."); } if (validateHeader(headers, recordMetaData.getFields())) { return; } throw new IOException("Invalid header."); }
private void addFieldBindings(Bean bean) { for (FieldMetaData fieldMetaData : vfRecordMetaData.getRecordMetaData().getFields()) { if (!fieldMetaData.ignore()) { bean.bindTo(fieldMetaData.getName(), recordElementName + "/" + fieldMetaData.getName()); } } }
@Override protected void validateHeader(List<String> headers) { // For backward compatibility with pre v1.5.... try { super.validateHeader(headers); } catch (IOException e) { throw new CSVHeaderValidationException(getFactory().getRecordMetaData().getFieldNames(), headers); } } }
/** * Get the unignored field count for the specified record. * * @param recordMetaData The record metadata. * @return The unignored field count. */ public int getUnignoredFieldCount(RecordMetaData recordMetaData) { if (factory.isMultiTypeRecordSet()) { // Need to account for the leading identifier field on each // record... return recordMetaData.getUnignoredFieldCount() + 1; } else { return recordMetaData.getUnignoredFieldCount(); } }
public CSVMapBinder(String fields, Class recordType, String keyField) { AssertArgument.isNotNullAndNotEmpty(fields, "fields"); AssertArgument.isNotNull(recordType, "recordType"); AssertArgument.isNotNullAndNotEmpty(keyField, "keyField"); smooks = new Smooks(); smooks.setFilterSettings(FilterSettings.DEFAULT_SAX); smooks.setReaderConfig(new CSVRecordParserConfigurator(fields) .setBinding(new Binding(beanId, recordType, BindingType.MAP).setKeyField(keyField))); }
/** * public constructor. * * @param name Record name. * @param fields Record fields metadata. * @param wildCardRecord Wildcard record. Accept any fields and generate the * field names based on index. */ public RecordMetaData(String name, List<FieldMetaData> fields, boolean wildCardRecord) { AssertArgument.isNotNullAndNotEmpty(name, "name"); AssertArgument.isNotNull(fields, "fields"); this.name = name.trim(); this.fields = fields; this.wildCardRecord = wildCardRecord; countIgnoredFields(); gatherFieldNames(); }
private void countIgnoredFields() { for (FieldMetaData field : fields) { if (field.ignore()) { ignoredFieldCount++; } else { unignoredFieldCount++; } } }
public FieldMetaData(String name) { AssertArgument.isNotNullAndNotEmpty(name, "name"); this.name = name; ignore = name.startsWith(IGNORE_FIELD); if (ignore) { ignoreCount = parseIgnoreFieldDirective(name); } }
public CSVListBinder(String fields, Class recordType) { AssertArgument.isNotNullAndNotEmpty(fields, "fields"); AssertArgument.isNotNull(recordType, "recordType"); smooks = new Smooks(); smooks.setFilterSettings(FilterSettings.DEFAULT_SAX); smooks.setReaderConfig(new CSVRecordParserConfigurator(fields) .setBinding(new Binding(beanId, recordType, BindingType.LIST))); }