Skip to main content

Migration notes

Migrate to 6.0-sb3

This release has been exceptionally released to provide a compatibility with Spring Boot 3 after the 6.0.0 release. When migrating from Spring Boot 2.x to Spring Boot 3.2, it brings a number of breaking changes, most of which are due to the upgrade to Spring Boot 3. Some notable changes independent of Atoti Server are the requirement for JDK 17, the removal of javax... related packages in favor of Jakarta EE, the upgrade to Spring 6.x with major changes to Spring Security. Or even the upgrade to Hibernate 6.x.

In addition to these changes, some updates had to be made in Atoti Server. These changes will be described below.

Migration from Spring Cloud Sleuth to OpenTelemetry

Spring Cloud Sleuth as defined with Spring 5.x does not exist with Spring 6.x. It was rebased on top of Micrometer. Instead of moving to Micrometer, which provides a Java-specific trace format, Atoti Server has moved to OpenTelemetry.

The configuration for setting up Tracing in projects needs to evolve. In particular, an instance of OpenTelemetry must be defined in applications and passed to Tracing#setOtelInstance manually or via the built-in configuration class TracingConfig.
See the how-to guide on Using OpenTelemetry in a project for detailed approach or the conceptual overview in here.

At the lowest level, all the plumbing evolving Tracer, Spans, SpanContexts, etc. has been changed from Sleuth classes to those of OpenTelemetry.
Because of this, the field com.qfs.msg.csv.impl.ChannelTask#threadPool changed to be an internal IExecutorService instead of a custom ParserThreadPool. Note that this ChannelTask class is highly internal and should not be overridden in projects.

ActiveMonitor custom Hibernate types

Hibernate 6 introduced the concept of Attribute Converter to store data in a custom format in Database columns. ActiveMonitor ships 3 special types to encode Event updated resources and states and PointValue coordinates. These attributes are now recorded using this new concept of Attribute converters, while maintaining the original format.

The classes com.qfs.repository.persistence.type.impl.StringCoordinateType, com.qfs.snl.persistence.type.impl.StringListType, com.qfs.snl.persistence.type.impl.StringStringMapType and com.qfs.snl.persistence.type.impl.StateListType have been removed. And their updated counterparts are now private classes.
The underlying class of com.qfs.snl.persistence.type.impl.AJsonConverter changed to become an attribute converter and became a private class.

Migrate to 6.0

danger

Starting from 6.0, many classes visible from Atoti jars are now considered internal classes.
These internal classes are located in packages containing internal or private_.
Examples:

  • com.activeviam.database.sql.internal.schema.ISqlFieldDescription, indicated by the "..sql.internal.schema..."
  • com.activeviam.database.bigquery.private_.BigqueryDataVersion, indicated by the "..bigquery.private_..."

These classes may be changed at any point in time, without regard for backward compatibility, without warning, without mention in the changelog, and without migration notes.
Users must not use them. This will never be an issue as they are never exposed to users The APIs always return their interfaces or wrap them. They are only visible through debuggers.

The main change in this version is the introduction of the Database API. This API will allow ActivePivot to work on top of the Datastore but also on top of external databases. The Datastore now implements the new Database API and the core components of ActivePivot also use it rather than the Datastore API.

Datastore

OperationBeforeNow
Compile a record querydatastore.getLatestVersion().getQueryManager().compile(query)datastore.getQueryManager().listQuery()
Compile a GetByKey queryDatastoreQueryHelper.createGetByKeyQuery()datastore.getQueryManager().getByKeyQuery()
Execute a querydatastore.getHead(branch).getQueryRunner().forQuery(query)datastore.getHead(branch).getQueryRunner().xxxQuery()
Get the metadatadatastore.getQueryMetadata().getMetadata()datastore.getEntityResolver()
  • Field dictionaries are no longer accessible (datastore.getQueryMetadata().getDictionaries() before) but you should never need them. Indeed, the new way to retrieve a field's default value (datastore.getEntityResolver().findTable(tableName).getField(fieldName).getType().getDefaultValue()) always returns an un-dictionarized value, as opposed as before when using recordFormat.getDefault(fieldIndex) that returns the dictionarized value when the field is dictionarized.
  • The ICursor interface now implements AutoCloseable. If cursors are used in your project, do not forget to close them, as they could leak database connections.
  • Methods to create conditions in BaseConditions now require arguments of the safe type FieldPath instead of legacy field expression strings. These methods have also been renamed to start with a lower case as they are not classes: Equal is now equal, And is now and... Note that Lesser and LesserOrEqual have been renamed less and lessOrEqual. True() and False() have been removed and can be replaced by BaseConditions.TRUE and BaseConditions.FALSE.
  • SelectionField is replaced by AliasedField. The equivalent of new SelectionField("alias", "reference/field") is now AliasedField.create("alias", FieldPath.of("reference", "field")).
  • The constructors of NonNullableFieldsPostProcessor and DictionarizeFieldsPostProcessor now use ReachableField instead of String.
  • The constructor of StoreDescriptionBuilder is now protected. Use StoreDescription.builder() instead.
  • In StoreDescriptionBuilder, all operations on fields should be done before defining the partitioning. For example, you cannot call addField() after withPartitioning().
  • The return type of DatastoreSchemaDescriptionUtil.createPath(String...) has changed. See the javadoc.
  • DynamicConditions was removed. Its methods are in BaseConditions. Sub-interfaces of ICondition and existing implementations have been moved and are now internal classes. ICustomCondition remains a public class, though it has been moved into com.activeviam.datastore.condition package.
  • The API to create Dynamic conditions has changed. It is now mandatory to name a parameter using as(String) method. Parameter indexes are no longer supported.
  • Other changes:
BeforeNow
datastoreVersion.getQueryRunner()datastoreVersion.getDatastoreQueryRunner()
datastoreVersion.getQueryManager()datastoreVersion.getDatastoreQueryManager()

Spring

The interface IActivePivotManagerDescriptionConfig no longer contains the datastore description. The IActivePivotManagerDescriptionConfig#userManagerDescription() was renamed managerDescription(). userSchemaDescription() was moved to and renamed IDatastoreSchemaDescriptionConfig#datastoreSchemaDescription(). One now has to define an IDatastoreSchemaDescriptionConfig that will contain the datastore description.

  • ActivePivotConfig was renamed ActivePivotWithDatastoreConfig.
  • IDatastoreConfig.datastore() was renamed IDatastoreConfig.database().

CXF & SOAP

Both SOAP services and CXF dependencies were completely dropped. Following this, all remote services require an AAuthenticator instead of a ClientPool.

Here is a list of replacements for each dropped service from com.qfs.webservices:

Webservice interface nameReplacement
IIdGenerator, IStreamingService, ILongPollingServiceActivePivot Websocket API
IQueriesServicecom.qfs.pivot.rest.query.IQueriesRestService
IAdministrationServicecom.qfs.pivot.rest.discovery.IDiscoveryRestService
IConfigurationServicecom.qfs.pivot.rest.configuration.IConfigurationRestService
ILicensingServiceNo replacement

ActivePivot

Aggregation module changes

Cloneable aggregated values

The API to use when writing aggregation bindings with cloneable aggregated values has been rewritten to avoid user mistakes.

Previously, users achieved such a task by extending ABasicAggregationBindingCloneable or one of its sub-classes - AAggregationBindingCloneable, AVectorAggregationBinding, SumVectorAggregationBinding, ... It often requires to carefully write aggregated values using #write(int, Object) in addition to an obscure flag using #writeBoolean. Reads and updates would also require reading such flags and values, not to mutate a value shared through the system.

The new API focuses on user ease. When writing a value, implementors must choose between #writeReadOnlyAggregate(int, Object), preserving the original value as much as possible or #writeWritableAggregate(int, Object), informing this class that the provided value can be mutated.
Similarly, when reading an existing aggregate, implementors can access the possibly untouched value through #readReadOnlyAggregate(int) or automatically access a writable value using #readWritableAggregate(int).
At any point, implementors are free to call the write methods to replace the current aggregates.

With the described changes, aggregating values change to something like this (in pseudo-code):

void copy(int from, int to){
// The aggregate is read-only because we pulled it directly from the input reader
// We avoid creating a copy of the value that may not be necessary
- this.output.write(to, this.input.read(from))
- this.output.writeBoolean(to, false);
+ writeReadOnlyAggregate(to, readInput(from));
}

void aggregate(int from, int to) {
- IVector aggregate = this.output.read(to);
- if (this.output.readBoolean(to)) {
- aggregate = cloneAggregate(aggregate);
- this.output.write(to, aggregate);
- this.output.writeBoolean(to, true);
- }
+ IVector aggregate = readWritableAggregate(to);
// We can safely modify the content of `aggregate`
aggregate.plus(readInput(from));
}
danger

When implementing an aggregation that uses AVectorAggregationBinding to bind columns, ensure that the IChunkFactory creates marked chunk, for instance writing:

public class AggregationVector extends AAggregation<?> {

@Override
public IChunkFactory<?> createChunkFactory(
final boolean isTransient,
final IAllocationSettings allocationSettings) {
return ChunkFactories
.chunkMarkedVectorFactory(getAggregatedDataType(), isTransient, allocationSettings);
}

}

The Aggregation Function documentation provides an up-to-date guide for writing this type of aggregation function.

Deprecated classes

The following abstract classes are deprecated and will be removed in a future version:

  • AGenericBaseAggregationFunction
  • AGenericAggregationFunction
  • AGenericVectorAggregationFunction and AVectorAggregationFunction (See example)

Implementers must now write their aggregation functions using the standard unique API. The Aggregation Function documentation provides an up-to-date guide for writing an aggregation function.

Generalization of Aggregation Functions

The client facing (through the Registry) IAggregationFunction has been extensively modified. @QuartetPluginValue(intf = IAggregationFunction.class) and @QuartetPluginValue(intf = IUserDefinedAggregationFunction.class) must be changed to @QuartetPluginValue(intf = IGenericAggregationFunction.class).

A new IGenericAggregationFunction has been introduced, which allows you to create an aggregation based on an arbitrary number of data sources. IAggregationFunction represents a specialization of this new interface, for the creation of aggregations based on a single source of data. We recommend clients implement AAggregationFunction when implementing aggregation functions based on a single source of data.

The interface IGenericAggregationFunction contains the following methods:

  • int getAggregatedType(int[] sourceTypes), which deduces the type of the aggregated values, based on the types of the sources of data. It must throw if the given types are not supported. It can be used to implement:
  • IGenericAggregation createAggregation(List<String> sourceIdentifiers, int[] sourceTypes), which creates an aggregation specialized for the given sources of data.
  • IGenericAggregationFunction withRemovalSupport(), which specializes the current aggregation function with capabilities to disaggregate from the resulting column.

The interface IAggregationFunction has lost the responsibility to create a chunk factory, which is now the prerogative of the IAggregation interface. IMultiSourceAggregationFunction was also introduced, with an associated base abstract class AMultiSourceAggregationFunction that mirrors AAggregationFunction.

IMultiSourceAggregation was also introduced, with an associated base abstract class AMultiSourceAggregation that mirrors AAggregation. We recommend clients implement AAggregation, AMultiSourceAggregation or AUserDefinedAggregation instead of implementing the interfaces, as they already give default implementations for most of the methods.

Post Processors

  • Legacy post processors have been replaced by their corresponding new implementation. This means all V2 post processors introduced in 5.11 have become the base post processors to use. Migration notes are available in post processor migration section.
  • Post Processors no longer need to implement Serializable.

Builders

  • com.activeviam.pivot.builders.StartBuilding was removed and is replaced by com.activeviam.builders.StartBuilding.

  • com.activeviam.builders.StartBuilding.entitlementsProvider() was moved to com.activeviam.pivot.security.builders.EntitlementsProvider.builder().

  • The withProperty(String, String) methods on the dimension, hierarchy and level builders where respectively renamed withDimensionProperty, withHierarchyProperty and withLevelProperty to avoid ambiguities.

Copper

Parts of the Copper API causing boxing were removed. Migration notes are available in copper migration section.

  • CopperStore.field(String) now takes a field name instead of an expression. If you used an expression, you must now use CopperStore.field(FieldPath).
  • CopperStore.withMapping(String, CopperLevel) now takes a field name instead of an expression. If you used an expression, you must now use CopperStore.withMapping(FieldPath, CopperLevel). Same for CopperStore.withMapping(String, String).
Testing API

The CubeTesterBuilderExtension provided for in the Copper Testing API now required a Supplier<CubeTesterBuilder> as first argument. The migration can be straightforwardly done by replacing the provided testerBuilder instance by the ()->testerBuilder supplier in your test classes.

Partial providers

  • The default implementation of IPartialProviderSelector.select has been removed, any custom implementation of IPartialProviderSelector must now implement this method.

Datastore Service

This service is now called the Database Service. The classes and interfaces related to the Datastore Service have been renamed. datastore was replaced by database to emphasize that this service works on any IDatabase. Here are some examples:

  • IDatastoreServiceConfiguration was renamed IDatabaseServiceConfiguration.
  • DatastoreRestServicesConfig was renamed DatabaseRestServicesConfig.
  • DatastoreService was renamed ReadOnlyDatabaseService.
  • DatastoreRestService was renamed DatabaseRestServiceController.
  • JsonDatastoreQueryResult was renamed JsonDatabaseQueryResult.
  • SecurityFromCubeToDatastoreFilterHook was removed.

The public vocabulary has also changed, with "database"/"table"/"join" being used instead of "datastore"/"store"/"reference". Examples of impacted classes:

  • IDatastoreFieldDescription was renamed IDatabaseFieldDescription
  • IDatastoreReferenceDescription was renamed IDatabaseJoinDescription
  • JsonDatastoreReference was renamed JsonDatabaseJoin
  • IStoreSecurity was renamed ITableSecurity
  • IStorePermission was renamed ITablePermissions
  • the endpoint datastore/data/stores was renamed database/data/tables
  • the endpoint datastore/data/storeNames was renamed database/data/tableNames
  • the endpoint datastore/discovery/references was renamed database/discovery/joins

Analysis Hierarchy

Constructor of AAnalysisHierarchy now requires the list of ILevelInfo of the corresponding hierarchy levels. In any cases, implementations of Java-based analysis hierarchy should be migrated to AAnalysisHierarchyV2. Either the description of the analysis must be given to the cube description, or an IAnalysisHierarchyDescriptionProvider plugin value should be created to go along with the custom Analysis Hierarchy. This provider description must have the same plugin key as the Analysis Hierarchy.

Introspecting Analysis Hierarchies

In order to define the introspecting levels of an analysis hierarchy, one must now set the selectionField of the levels of the created AnalysisHierarchyDescription according to the needs of the hierarchy.

Analysis hierarchies with at least now one level with a specified selectionField in their description are now implicitly introspecting hierarchies and must implement AMultiVersionAnalysisHierarchy#processIntrospectedMember([...]) to define their introspection logic.

Limits regarding the ordering between introspecting and non-introspecting levels for an Analysis Hierarchy are unchanged (impossible to define an introspecting level deeper than a non-introspecting one).

Hierarchy descriptions

Hierarchy descriptions now must use the IHierarchyDescription interface. Depending on the type of described hierarchy, the sub-interface to use will change:

  • "Standard" hierarchy descriptions (e.g. hierarchies whose levels are filled by levels of the Selection) use the IAxisHierarchyDescription interface which remains unchanged.
  • All Analysis hierarchy descriptions now use the IAnalysisHierarchyDescription interface and now must explicitly contain information relative to their levels.
  • The fluent builder path for Analysis hierarchies has changed, from
withHierarchy(hierarchyName)
.withPluginKey(pluginKey)
.withProperty(propertyKey, propertyValue)

to :

withAnalysisHierarchy(IAnalysisHierarchyDescription)

or

withAnalysisHierarchy(hierarchyName, pluginKey)
.withCustomizations(hierarchyDescription -> {
hierarchyDescription.putProperty(propertyKey, propertyValue);
})

The second option allows for customization of the created Analysis Hierarchy's description. This description will be created by the associated IAnalysisHierarchyDescriptionProvider.

  • AAnalysisHierarchy.buildDiscriminatorPaths() was removed. It is replaced by a more efficient method: buildDiscriminatorPathsIterator(IDatabaseVersion).

-AxisLevelDescription(String levelName) was removed. A selectionField (formerly propertyName) must now be specified in the constructor to avoid unexpected behavior.

Aggregation Procedure

DatastorePrefetchRequest is replaced by DatabasePrefetchRequest. The field expressions are now represented with the FieldPath class instead of String.

Memory Allocation Monitoring

We are no longer reporting the off-heap memory allocated with java's internal MBeans. Instead, you can use the newly created MBeans located in DirectMemoryMonitoring for the information you wanted regarding memory consumption from ActivePivot. Additionally, MAC can be used for deeper investigations.

Other changes

  • StartBuilding.selection(IDatastoreSchemaDescription) was removed. Use StartBuilding.selection(description..asDatabaseSchema()) instead.
  • ADatastoreVersionAwareProperty was renamed ADatabaseVersionAwareProperty.
  • com.qfs.store.IDatastoreVersionAware was renamed com.activeviam.database.api.IDatabaseVersionAware.
  • IDatastoreSelectionSession was renamed ISchemaSelectionSession.
  • IFilteredDatastoreSelectionSession was renamed IActivePivotSchemaSelectionSession.
  • IDatastoreAware was changed and will be replaced by IDatabaseAware.
  • AStoreStream.dictionaries is replaced with AStoreStream.resultFormat.getDictionary(...).
  • All ActivePivot named threads are now prefixed by activeviam and suffixed by worker-i where i is the thread count number.
  • Creating a QueriesTimeLimit with a Duration lower than 1 second will now throw a IllegalArgumentException.

Content Server Rest API

  • The new Rest Bulk API for Content Server was developed. The old API for bulk operations was removed. It is no longer possible to use POST method with query parameters moveAll, putAll or deleteAll because mappings for them are deleted. Also, the query parameter metadataOnly is replaced by a parameter includeContent that is a logical opposite of metadataOnly. Do not forget to update the value of this parameter from True to False and vice versa when renaming.

CSV Source

  • The class CSVSourceConfiguration is now created through CSVSourceConfigurationBuilder, and is given the method fileComparator that previously belonged to the class CSVSource.

Rest Services

All our Rest endpoints were migrated to Spring MVC. The migration implies a few changes:

  • Removal of the JsonResponse wrapper. The data will now be directly in the body of the request and the status in the HTTP Status code of the request. The errors in AP are still returned as JsonError with the full stack trace

  • Every rest endpoint is now prefaced by the activeviam namespace. For example : localhost:XXXX/activeviam/content/rest/v7/

  • The datastore endpoint has been changed to database.

  • The content server rest service now handle the path of the file requested or given through Query parameters not Path variables. The parameter for import, permissions, etc. needs now to be true and doesn't allow void.

  • The Activemonitor's monitors site separator is now an underscore not a slash.

ActiveMonitor

  • IParameterAwareActivePivotManagerDescriptionConfig was renamed IParameterAwareDatastoreSchemaDescriptionConfig. You will need to add a Spring configuration implementing IActivePivotManagerDescription in your project.

Properties

  • The properties used for the external configuration of ActivePivot and ActiveMonitor applications in ActiveViamProperty and JwtConfig have been changed to uniformize their naming conventions. Properties are now either :
    • Prefixed by "activeviam"
    • Prefixed by an ActiveViam technical product name such as "activepivot" or "activemonitor"

List of the affected properties with the new corresponding naming :

  • In JwtConfig :
BeforeNow
qfs.jwt.expirationactiveviam.jwt.expiration
qfs.jwt.key.privateactiveviam.jwt.key.private
qfs.jwt.key.publicactiveviam.jwt.key.public
qfs.jwt.claim_key.authoritiesactiveviam.jwt.claim_key.authorities
qfs.jwt.claim_key.principalactiveviam.jwt.claim_key.principal
qfs.jwt.generateactiveviam.jwt.generate
qfs.jwt.check.user_detailsactiveviam.jwt.check.user_details
  • In ActiveViamProperty :
BeforeNow
contentServer.remote.import.timeoutactiveviam.contentServer.remote.import.timeout
qfs.handler.stored.maxSizeactiveviam.handler.stored.maxSize
qfs.mmap.thresholdactiveviam.mmap.threshold
qfs.mmap.trackingactiveviam.mmap.tracking
qfs.compression.chunkset.disableactiveviam.compression.chunkset.disable
qfs.stream.configuration.gc.delayactiveviam.stream.configuration.gc.delay
qfs.mdx.formula.callStackDepthLimitactiveviam.mdx.formula.callStackDepthLimit
completer.log.suppressedactiveviam.completer.log.suppressed
qfs.contentservice.rootactiveviam.contentservice.root
nativeMemoryCacheRatioactiveviam.nativeMemoryCacheRatio
qfs.slab.memory.allocation.strategyactiveviam.slab.memory.allocation.strategy
defaultChunkSizeactiveviam.defaultChunkSize
minimalChunkSizeactiveviam.minimalChunkSize
chunkGarbageCollectionFactoractiveviam.chunkGarbageCollectionFactor
chunkAllocatorClassactiveviam.chunkAllocatorClass
qfs.vectors.defaultBlockSizeactiveviam.vectors.defaultBlockSize
qfs.vectors.garbageCollectionFactoractiveviam.vectors.garbageCollectionFactor
qfs.vectors.allocator.pool.sizeactiveviam.vectors.allocator.pool.size
qfs.vectors.swap.directory.numberFilesactiveviam.vectors.swap.directory.numberFiles
vectorDelimiteractiveviam.vectorDelimiter
queries.continuous.primitive.fullForkactiveviam.queries.continuous.primitive.fullFork
postprocessor.partitionOnRangeLevelsByDefaultactiveviam.postprocessor.partitionOnRangeLevelsByDefault
qfs.distribution.gossip.router.portactiveviam.distribution.gossip.router.port
qfs.distribution.gossip.router.enableactiveviam.distribution.gossip.router.enable
logicalAddressactiveviam.logicalAddress
protocolPathactiveviam.protocolPath
qfs.distribution.netty.externalAddressactiveviam.distribution.netty.externalAddress
qfs.distribution.netty.portRangeactiveviam.distribution.netty.portRange
nettyMessageMaxSizeactiveviam.distribution.nettyMessageMaxSize
qfs.distribution.netty.bindAddressactiveviam.distribution.netty.bindAddress
qfs.streaming.MDX.attemptLimitactiveviam.streaming.MDX.attemptLimit
qfs.streaming.GET_AGGREGATES.attemptLimitactiveviam.streaming.GET_AGGREGATES.attemptLimit
xmlaDiscoveryApproxCardinalityactiveviam.xmlaDiscoveryApproxCardinality
qfs.pool.policyactiveviam.pool.policy
qfs.pool.sizeactiveviam.pool.size
qfs.pool.scheduler.sizeactiveviam.pool.scheduler.size
quartet.property.separatoractiveviam.property.separator
qfs.collectible.query.typeactiveviam.collectible.query.type
qfs.mmap.sampling.depthactiveviam.mmap.sampling.depth
qfs.mmap.sampling.startactiveviam.mmap.sampling.start
qfs.mmap.sampling.percentactiveviam.mmap.sampling.percent
qfs.branch.masteractiveviam.branch.master
qfs.pool.nodesactiveviam.pool.nodes
qfs.pool.procsactiveviam.pool.procs
qfs.activepivot.expiredqueries.pollingactiveviam.activepivot.expiredqueries.polling
qfs.streaming.initialPublicationModeactiveviam.streaming.initialPublicationMode
qfs.exportservice.rootpathactiveviam.exportservice.rootpath
duringThreadNumberactiveviam.duringThreadNumber
maxDuringThreadNumberactiveviam.maxDuringThreadNumber
qfs.distribution.remote.pool.sizeactiveviam.distribution.remote.pool.size
qfs.distribution.log.size.thresholdactiveviam.distribution.log.size.threshold
qfs.distribution.maxPendingDiscoveriesactiveviam.distribution.maxPendingDiscoveries
qfs.conflation.maxQueueSizeactiveviam.conflation.maxQueueSize
qfs.conflation.maxDelayTimeactiveviam.conflation.maxDelayTime
repository.cache.isolated_transactionsactiveviam.repository.cache.isolated_transactions
repository.poll.periodactiveviam.repository.poll.period
repository.poll.period.maxactiveviam.repository.poll.period.max
repository.poll.log.thresholdactiveviam.repository.poll.log.threshold
repository.daemon.waitStableDistributionactiveviam.repository.daemon.waitStableDistribution
qfs.server.namespace.parentNO LONGER USED (see Rest Services section)
continuousQueryMonitoringactiveviam.continuousQueryMonitoring
qfs.selection.listener.catchUpMaxTimeactiveviam.selection.listener.catchUpMaxTime
com.activeviam.json.strongPrimitiveParsingactiveviam.json.strongPrimitiveParsing
com.activeviam.directquery.enableAutoVectorizeractiveviam.directquery.enableAutoVectorizer
com.activeviam.directquery.cubeFeedingTimeoutInSecondsactiveviam.directquery.cubeFeedingTimeoutInSeconds
com.activeviam.directquery.snowflake.maxresultsetsizeactiveviam.directquery.snowflake.maxresultsetsize
mdx.negativecellset.limitactiveviam.mdx.negativecellset.limit
qfs.activecollector.renew.frequencyactiveviam.activecollector.renew.frequency
activepivot.snl.urlactivemonitor.activepivot.url
live.snl.urlactivemonitor.ui.url
sentinel.poll.periodactivemonitor.poll.period
sentinel.poll.period.maxactivemonitor.poll.period.max
sentinel.poll.log.thresholdactivemonitor.poll.log.threshold
sentinel.daemon.waitStableDistributionactivemonitor.daemon.waitStableDistribution
sentinel.periodicExecutors.poolSizeactivemonitor.periodicExecutors.poolSize

Migrate to 5.11

Spring

Spring was updated to version 2.6. As a result of the necessary configuration changes, the class com.activeviam.cfg.SpringRestServiceInitializer must be added to your Spring configuration, unless you already import com.qfs.server.cfg.impl.ActiveViamRestServicesConfig or com.qfs.content.cfg.impl.StandaloneContentServerRestConfig to configure the REST services.

Copper

The performances of the calculations defined in Copper have been improved. The transient memory generated at query time will be reduced by limiting the boxing phenomenon. (The use, for instance, of java.lang.Integer instead of int)

As a consequence, multiple methods are now deprecated. We strongly advise users to migrate out of any deprecated methods, as these changes are strictly brought to users for transient memory performance purposes.

  • Combining multiple CopperElements through Copper.combine(...) to create a new measure is no longer done with CopperMeasure map(SerializableFunction<IRecordReader, Object> mapper). The method is now called mapToObject to follow the changes brought to other methods. For instance,

    Copper.combine(Copper.sum(RISK__PNL), Copper.member(City))
    .mapToInt(reader -> {
    if (reader.read(1).equals("NYC") {
    return 0;
    } else {
    return reader.readDouble(0) > 0 ? 1 : -1;
    }
    })
  • Combining multiple CopperMeasure through Copper.combine(...) to create a new measure is no longer done with CopperMeasure map(SerializableFunction<IRecordReader, Object> mapper). Instead, the new method's output type will dictate the method that should be used. For instance, if these measures are combined into a measure that outputs int, one should use CopperMeasure mapToInt(SerializableToIntFunction<IArrayReader> mapper). For instance,

    Copper.combine(Copper.sum(RISK__PNL), Copper.constant(0d).withName(PNL_LIMIT))
    .mapToInt(a -> a.readDouble(0) >= a.readDouble(1) ? 1 : -1)
  • Modifying a measure follows the same rule : Copper.sum(RISK__PNL).mapToDouble(reader -> reader.readDouble(0) * 2d)

  • The aforementioned changes make CopperMeasure#withType() relatively redundant: it is removed for measures that have already specified it through the choice of mapping method.

  • Contrary to the specialized implementations like mapToInt(SerializableToIntFunction<IArrayReader>), mapToDouble(SerializableToDoubleFunction<IArrayReader>, etc... The method mapToObject takes as argument a SerializableBiConsumer<IArrayReader, IWritableCell>. The provided cell must be used to write the result:

    Copper.sum(RISK__PNL).mapToObject((reader, cell) -> {
    if (reader.readDouble(0) <= 1) {
    cell.writeDouble(reader.readDouble(0));
    }
    });
  • From now on, only two types of measures support setting a specific output type: when using the various mapToDouble/mapToLong methods and the new map, or right after a plus/minus/etc. using withType.

Post Processors

To also reduce the memory consumption of IPostProcessors and IEvaluators, we added new implementations of these components.

AAdvancedPostProcessor<outputType> becomes AAdvancedPostProcessorV2 and does not declare a generic output type. All Post Processors must now receive the property outputType in their initialization properties.

  • IEvaluatorV2's evaluation method is void evaluate(ILocation location, IRecordReader aggregatedMeasures, IWritableCell resultCell). Instead of reading underlying values from an Object[], one now needs to read them from the given IRecordReader, using the appropriate specialized method, and write the result of the evaluation into the resultCell. For instance,

    @Override
    public Double evaluate(ILocation location, Object[] measures) {
    if (measures[0] == null) {
    return null;
    } else {
    return 10D + ((Number) measures[0]).doubleValue();
    }
    }

    becomes

    @Override
    public void evaluate(ILocation location, IRecordReader aggregatedMeasures, IWritableCell resultCell) {
    if (!aggregatedMeasures.isNull(0)) {
    resultCell.writeDouble(10D + aggregatedMeasures.readDouble(0));
    }
    }
  • In a similar fashion, all abstract implementations of ITransformProcedure are now deprecated. It is now necessary to directly implement the interface. As before, one can now either implement the method transform or the method doTransform, depending on the expected behavior.

  • Similarly, the method evaluateLeaf of the ADynamicAggregationPostProcessor was reworked to provide a cell in ADynamicAggregationPostProcessorV2

  • The method IPostProcessor#init(Properties) has been completely reworked, with clear methods to implement or override for each of the most important parts of the post processor.

Distribution

Now, cubes within a distributed cluster have a unique identifier that will be used to expose their remote addresses through the query cube they are related to. This introduces a breaking change in the distributed messenger constructors: ADistributedMessenger now include CubeEndPointInfo in its constructor, a data structure holding information regarding cube addresses and identifiers. These data are solely resolved by core ActivePivot, thus one has just to include the new field in any subsequent messenger.

Analysis Hierarchies

The implementations of Analysis Hierarchies have been cleaned as part of 5.11. Users are expected to extend and override AAnalysisHierarchyV2.
Most users should not see a difference because Analysis Hierarchies were mostly populated with static members.

The contract between the two classes has been better described. There is now a clear separation between static members and members generated from the ActivePivot Schema.
Static members are still created by AAnalysisHierarchyV2#buildDiscriminatorPaths.
However, to process introspected members, the new method is #processIntrospectedMember, replacing a judicious override of #contributeMember. Instead of consuming a member and having to call the super method to do the contribution, users receive the introspected members and their counts, as well a consumer function. It is up to the function to generate as many members as it wants and to pass each one of them to the consumer to publish them.

As before, AAnalysisHierarchyV2 sources are available for inspection and inspiration.

Datastore

  • To create datastore queries one should now give a SelectionField instead of a String which was the path to the field.
  • ReferencedField was renamed ReachableField.
  • ISelectionField was removed. It is replaced by SelectionField.
  • SelectionField.getName() was renamed getAlias().

getMostRecentVersion() should never be used by a project. It is an internal method. Replace all calls to getMostRecentVersion() by a call to getHead().

Vectors

All methods in IVectorBinding were changed from: void methodName(IVector source, IVector destination) to: void methodName(IVector left, IVector right)

This change is performed to align all the vector APIs together. For instance, doing vectorA.minus(vectorB) is now, under the hood, calling binding.minus(vectorA, vectorB), keeping the arguments in their initial order. This should better prevent mistakes in the implementations of methods such as IVectorBinding.applyAsDouble(left, right, DoubleBinaryOperator).

Users writing their own vector implementations and their own bindings for these vectors will probably need to ensure their bindings are in order.

Content Service UI

ContentServerResourceServerConfig is replaced by AdminUIResourceServerConfig and content-server-ui.jar is replaced by admin-ui.jar.

Content Server and ActiveMonitor Database

Migrating H2 to v2

The dependency version of h2 for ActivePivot and ActiveMonitor was upgraded from 1.4.200 to 2.1.214. This version change makes h2-generated .db files incompatible between 6.0 and the prior versions.

Unwanted .db files can simply be deleted, or restoring the entries of your current database can be attempted using the h2 migration notes.

SOAP

We no longer return a full ActivePivotManagerDescription in our webservices with retrieveManagerDescription but instead return a lighter description called SoapActivePivotDiscovery in retrieveActivePivotDiscovery. This discovery will contain all the information you previously wanted from an ActivePivotManagerDescription including a list of Catalogs and a list of Cubes (And in each cube, its measures, dimensions, hierarchies and levels).

retrieveActivePivotDescription also no longer return a IActivePivotDescription but a SoapCubeDiscovery which is a lighter description of your ActivePivot.

retrieveSchemaDescription no longer exists in our webservices.

XML

Building ActivePivot from an XML description file is no longer available. Users are invited to use our fluent builder API using StartBuilding.

ActiveviamProperties

The options for the ActiveViam Property -Dactiveviam.chunkAllocatorClass were moved to specialized packages. See the Javadoc of the property CHUNK_ALLOCATOR_CLASS_PROPERTY.

Migrate to 5.10

Datastore

  • Primary indexes are now called Unique indexes (All the rows in an index of this type must be unique).
  • TransactionManagerUtil.resetBranch was moved in ITransactionManager.
  • Rename ARemoveUnknowKeyListener to ARemoveUnknownKeyListener to fix typo.
  • Move NumaSelectorFromStoreDescription from com.qfs.desc to com.qfs.desc.impl package.
  • Move ADefaultEpochPolicy from com.qfs.multiversion to com.qfs.multiversion.impl package.
  • Rename ImmutableRecordList to ImmutableRecordBlock to match the naming pattern used by sibling classes.

ActivePivot

  • Change the hierarchy builder: hierarchyBuilder.factless().withStoreName(...) was replaced by hierarchyBuilder.fromStore(...). It cannot be called after factless() nor notFactless().
  • Rename MultiVersionAxisHierarchy.FIELD_NAME_PROPERTY to MultiVersionAxisHierarchy.FIELD_EXPRESSION_PROPERTY.

Copper

  • All published Copper measure must now be named. Expressions like Copper.sum("pnl").publish(context) have to be replaced with Copper.sum("pnl").withName("pnl.SUM").publish(context).
  • Copper.newSingleLevelHierarchy(...).from(CopperStoreField) was replaced by Copper.newHierarchy(...).fromField(CopperStoreField).
  • Copper.newSingleLevelHierarchy(...).from(CopperLevelValues) was replaced by Copper.newHierarchy(...).fromValues(CopperLevelValues).

Distribution

  • The JGroupsMessenger has been removed. Now, ActivePivot relies on two messenger types, the LocalMessenger and the NettyMessenger. Note that the NettyMessenger still uses JGroups for member discovery and group membership. More details on this topic in Distributed Messenger
  • The property NotificationMember#AWAIT_NOTIFICATIONS has been renamed to IDataClusterDefinition#AWAIT_NOTIFICATIONS_TIME.

Functional Interfaces

The functional interfaces in package com.qfs.func have been moved and renamed.

Old locationNew location
com.qfs.func.IBiIntPredicatecom.activeviam.util.function.BiIntPredicate
com.qfs.func.IFloatUnaryOperatorcom.activeviam.util.function.FloatUnaryOperator
com.qfs.func.IntBiFunctioncom.activeviam.util.function.IntBiFunction
com.qfs.func.ITriFunctioncom.activeviam.util.function.TriFunction
com.quartetfs.fwk.util.FloatBinaryOperatorcom.activeviam.util.function.FloatBinaryOperator
com.qfs.func.IEithercom.activeviam.util.function.IEither
com.qfs.func.IOptionalEithercom.activeviam.util.function.IOptionalEither
com.qfs.func.impl.Eithercom.activeviam.util.function.impl.Either
com.qfs.func.impl.OptionalEithercom.activeviam.util.function.impl.OptionalEither

Query Results Context Value

The default behavior for QueriesResultLimit context value have changed from QueriesResultLimit#withoutLimit() to QueriesResultLimit#defaultLimit() (the default transient and intermediate result limit amount to 100,000 and 1,000,000 point locations respectively). The default limits may appear low for certain projects, in this case the configuration can be easily overridden. This is explained in the respective documentation section.

ActiveViam Properties

  • activeviam.datastore.query.maxLookupsOnPrimary was renamed activeviam.datastore.query.maxLookupsOnUnique.
  • ActiveViamProperty.MAX_LOOKUPS_ON_PRIMARY_INDEX was renamed ActiveViamProperty.MAX_LOOKUPS_ON_UNIQUE_INDEX.

Spring Configuration Changes

The ActivePivot Configuration does not automatically contain the ContextValueFilter bean anymore, thus preventing conflicts with Spring Boot.

Azure Cloud Source migration

The Azure Cloud Source has migrated its dependency to the Azure Blob Storage SDK from v8 to v12.

A migration guide from v8 to v12 can be found on the Azure SDK repository.

Renamed Classes

All classes of the Azure Cloud Source have been changed to include the word Azure as a prefix. As a result, the following classes had their name changed:

5.95.10
CloudBlobPathAzureBlobPath
BlockBlobPathAzureBlockBlobPath
AppendBlobPathAzureAppendBlobPath
PageBlobPathAzurePageBlobPath

Changed Classes

  • The handling of client-side encryption has undergone some major changes to better match the new architecture of the Azure Blob Storage SDK. Changes about this are detailed in the below section.

  • IAzureCloudDirectory:

    • Now extends ICloudDirectory<BlobClientBase> instead of ICloudDirectory<CloudBlob>.
    • getTBlob(String) return types were updated with the new types from the Azure SDK.
  • Removed IAzureEntityPath.upload(InputStream, long, BlobRequestOptions).

    Prefer using getUnderlying() to get a reference to the underlying blob client and use blob-specific configuration for uploads.

  • Uploading with an unknown length is not supported anymore for page blobs, and will throw an UnsupportedOperationException if attempted.

  • Records#getGlobalDefaultValue directly uses the content type value passed as an int instead of ILiteralType.

Client-side Encryption

In the Azure Blob Storage SDK, client-side encryption is now implemented in the package com.azure:azure-storage-blob-cryptography.

With the SDK upgrade, client-side encryption cannot be configured at the Azure service client level anymore. It needs to be specified on a per-blob basis, before uploading or downloading content.

As a result, two new classes have been introduced to handle client-side encryption:

  • AzureEncryptedCloudDirectory: a variant of AzureCloudDirectory that additionally holds encryption keys for client-side encryption, and that can produce AzureEncryptedBlobPaths to existing and non-existing encrypted blobs.
  • AzureEncryptedBlobPath: a path to an encrypted blob; essentially a wrapper around an EncryptedBlobClient.

The other classes AzureBlobPath, AzureBlockBlobPath, AzureAppendBlobPath, AzurePageBlobPath and AzureCloudDirectory are all unaware of client-side encryption. They will not encrypt uploaded data and will not decrypt downloaded data.

They accept three additional arguments in their constructor compared to their non-encrypted counterparts:

  • a key-wrapping algorithm (String)
  • a key encryption key (com.azure.core.cryptography.AsyncKeyEncryptionKey)
  • a key encryption key resolver (com.azure.core.cryptography.AsyncKeyEncryptionKeyResolver)

If the directory or path only needs to perform either only downloads or only uploads, some arguments are not required and can be set to null: | operations | required | |---------------------|----------------------------------------------| | download | key encryption key resolver | | upload | key-wrapping algorithm, key encryption key | | download and upload | all arguments |

The possible key wrapping algorithms are specified in the class KeyWrapAlgorithm from com.azure:azure-security-keyvault-keys (the dependency is not included in the Azure Cloud Source).

Some snippets of AsyncKeyEncryptionKey creation from a local key or from an Azure Key Vault can be found in the readme of com.azure:azure-storage-blob-cryptography.

Important: the new implementation of client-side encryption in Azure Blob Storage SDK v12 only permits uploads to client-side-encrypted block blobs. The creation of client-side-encrypted blobs of other types (append and page) is not supported anymore.

Downloads of client-side encrypted page and append blobs that were already created through other means is still possible through the AzureEncryptedBlobPath.

Migrate to 5.9

Java 8 is no longer supported. ActivePivot 5.9 is compatible with Java 11.

ActivePivot

New Copper API

Shipped as an experimental feature in ActivePivot 5.8.3, the new Copper API is officially replacing the one released with ActivePivot 5.7.0, which is now abandoned.

Shift measure creation

The Copper shift measure creation methods have changed : CopperMeasure#at([...]) methods have been changed to a unified CopperMeasure#shift(CopperLevelsAt...) method. Find more information in our user guide for updated examples.

Parent Value

The method Copper.parentValue(measure, hierarchy) has been changed to measure.parentValueOn(hierarchy).

Setting the designated type of measure

To improve clarity, the method CopperMeasure#cast([...]) has been renamed to CopperMeasure#withType([...]).

Level and hierarchy representation

LevelCoordinate has been renamed and is now LevelIdentifier. HierarchyCoordinate has been renamed and is now HierarchyIdentifier.

Test jar

The Copper test utils that were previously packaged in a test-jar have been moved to their own module:

 <groupId>com.activeviam.activepivot</groupId>
<artifactId>activepivot-copper-test</artifactId>

Spring Configuration Changes

Schema configuration

ActivePivotConfig now requires an IActivePivotManagerDescriptionConfig. IDatastoreDescriptionConfig was merged into IActivePivotManagerDescriptionConfig and IParameterAwareDatastoreDescriptionConfig was renamed IParameterAwareActivePivotManagerDescriptionConfig.

In IActivePivotManagerDescriptionConfig you must now implement userManagerDescription() (respectively named userSchemaDescription()) instead of managerDescription() (respectively schemaDescription()).

The way the description post-processors are applied has been changed. You are impacted if you explicitly build the Datastore or the ActivePivotManager in your project instead of using DatastoreConfig and ActivePivotConfig. In this case, define an IActivePivotManagerDescriptionConfig and give the results of the managerDescription() and schemaDescription() default methods to the builders.

IActivePivotConfig.activePivotManagerDescription() was removed and replaced by IActivePivotManagerDescriptionConfig.managerDescription().

CORS configuration

The CORS configuration now relies on Spring standards. Instead of our own ICorsFilterConfig that created a SpringCorsFilter, we moved to Spring CorsConfigurationSource. As internal components require a knowledge of the CORS configuration before we can define the actual CorsConfigurationSource, a new interface ICorsConfig has been introduced, detailing the information of a standard CORS configuration. This is used in internal services to create REST and WS services compatible with the configuration. This is used in the sandbox project to create the final CORS configuration.

ICorsConfig only requires you to define the list of allowed origins. The other methods provide the accepted and exposed headers the authorized methods. Those have default implementations compatible with the ActivePivot stack.

SpringCorsFilter has been deleted. Thanks to CorsConfigurationSource, Spring automatically creates the filter.

Miscellaneous

ActivePivotServicesConfig removes an internal attribute, creating an unused dependency to IDatastoreConfig. If needed, you can restore this dependency by extending the configuration class and restoring the attribute.

I18nConfig was renamed AI18nConfig. Many of the static methods have changed to instance methods to allow easier extensions and refinements.

Aggregation Procedures

Interface

The interface was slightly changed to support more features in Copper (getDatastorePrefetch and createContext). This change should be transparent for procedures extending AAnalysisAggregationProcedure.

Additional validation checks

To offer support for Aggregation Procedures depending on other Procedures, ActivePivot uses the property IAnalysisAggregationProcedureDescription#getUnderlyingLevels() to identify the dependencies between Procedures. This drives a change to the API, preventing Procedures to define their own handled levels in the list of underlying levels.

This constraint is enforced by the validation procedure of the description. Configurations may need to be updated. Particularly, in 5.8, you always had to define a handled hierarchy from builders. You can now avoid this with the method withoutHierarchies().

Option to disable the Epoch level

The epoch level of the epoch dimension can now be enabled or disabled. The epoch dimension is still enabled by default. The epoch level is now disabled by default. To enable it, use IEpochDimensionDescription.setEpochLevelEnabled(boolean) or IEpochDimensionBuilder.withEpochLevel.

The IEpochDimensionBuilder.withBranchesLevel method was renamed IEpochDimensionBuilder.withBranchLevel. The methods in IEpochDimensionDescription were also renamed from *epochsLevel* and *branchesLevel* (e.g. getEpochsLevelComparator or getBranchesLevelFormatter) to *epochLevel* and *branchLevel* (e.g. getEpochLevelComparator and getBranchLevelFormatter).

Distributed Cube no longer supports the epoch level

New default aggregation functions

Aggregation functions that maintain a history state, like AVG, MIN, or MAX are memory intensive. Thus, for the MIN and MAX aggregation functions, the new default behavior is append-only. This means that disaggregation is not supported for these two functions. For disaggregation support, please use MIN_HISTORY and MAX_HISTORY aggregation functions (e.g. value.MIN_HISTORY or value.MAX_HISTORY).

Default Distributed Message Size Limits

Previous ActivePivot versions had a maximum default distributed message size of 1GB for every message type. This limit has been reduced. The default maximum size is:

  • 64MB for IInitialDiscoveryMessage messages
  • 4MB for ITransactionCommittedMessage messages
  • 4MB for ScopedAggregatesRetrievalResultTransporter answers wrapped in BroadcastResult messages
  • 32MB for DrillthroughMessageWithHeadersAnswers answers wrapped in BroadcastResult messages
  • 2MB for any other message types

You can still override these values using the property NETTY_MAX_SIZE_PROPERTY of the ActiveViamProperty. In addition to the older configuration method using class names, it is possible to configure the message default size limit using the constants (enum) defined in NettyStreamUtils#MessageType as follows:

  • INITIAL_DISCOVERY_MESSAGE for initial discovery messages
  • ITRANSACTION_COMMITTED_MESSAGE for transaction commit message
  • SCOPEDAGGREGATES_RETRIEVAL_TRANSPORTER for single Gaq query result messages
  • DRILLTHROUGH_HEADER for drillthrough header query result messages
  • DRILLTHROUGH_MESSAGE_WITH_HEADERS_ANSWER for drillthrough query result messages
  • GLOBAL_MESSAGE for global configuration

For instance, the following statement used in prior versions:

- -Dqfs.distribution.netty.message.maxSize=com.quartetfs.biz.pivot.cellset.impl.ScopedAggregatesRetrievalResultTransporter=42m,com.qfs.messenger.message.IInitialDiscoveryMessage=200m

becomes:

+ -Dactiveviam.distribution.nettyMessageMaxSize=com.quartetfs.biz.pivot.cellset.impl.ScopedAggregatesRetrievalResultTransporter=42m,com.activeviam.messenger.message.IInitialDiscoveryMessage=200m
or
+ -Dactiveviam.distribution.nettyMessageMaxSize=singleGaqResult=42m,initialDiscovery=200m

Miscellaneous

  • ALocationShiftPostProcessor#handleNoUnderlyingMeasure is renamed into ALocationShiftPostProcessor#handleNoUnderlyingMeasure to fix the typo - missing "g".

REST and WS APIs

With the introduction of a new REST call to export query plans and another REST endpoint to forward client traces to a tracing server, the version of ActivePivot REST API has changed from v4 to v5. With the addition of the metadata to the WS updates of the Content Service entries, the WS API has changed from v4 to v5.

As components of the REST and WS API versions remain synchronized, these changes result in the following URL changes:

BeforeAfter
pivot/rest/v4/...pivot/rest/v5/...
pivot/ws/v4/...pivot/ws/v5/...
content/rest/v4/...content/rest/v5/...
content/ws/v4/...content/ws/v5/...

Impacted REST services:

  • ActivePivot services for queries, context configurations, etc
  • Datastore service for queries and updates
  • Content Service
  • Tracing Service

CSV Source

The CSV Source can now accept incomplete lines (where the number of columns is smaller than the expected number of attributes of a record). You can set this in CSVParserConfiguration's full constructor, or through its setAcceptIncompleteLines(boolean). It can finally be set through ParserContext#setAcceptIncompleteLines(boolean) method. This sets this property on its underlying CSVParserConfiguration.

You can also sample the input files to work on a small portion of the input files. The sampling policy is passed as an argument of type ICsvParserConfigPolicy to the constructor of CSVParserConfiguration, or using the method CSVParserConfiguration#setParserPolicy(ICsvParserConfigPolicy).
The Core Product ships basic policies to load the first lines of one or more files, or load a limited number of files. Those are available in ICsvParserConfigPolicy static methods.

Parquet Source

The Parquet Source benefited from the same sampling policies as the CSV Source. They implement IParquetReaderPolicy, that are passed in the constructor of ParquetParser. Like the CSV Source, basic policies are available in IParquetReaderPolicy to load a limited number of lines or files.

Other changes

  • LogWriteException is now a runtime exception. This exception is still thrown when an error occurs during the writing of the Transaction Log.
  • One of the signatures of ITransactionManager.performInTransaction was removed.

Migrate to 5.8

Announcement

JDK 11 Support

ActivePivot 5.8 supports both JDK 8 and JDK 11. The Java version used by Maven when compiling ActivePivot depends on the JAVA_HOME environment variable, or the current java version if the JAVA_HOME environment variable is not set. For more information, see the dedicated page.

ActivePivot

HierarchyUtil

The following methods do not return null but throw an UnknownOlapElementRuntimeException when the element is not found:

public static IHierarchy getHierarchy(IActivePivot pivot, String hierarchyDescription) {...}

public static IHierarchy getHierarchy(IActivePivot pivot, String dimensionName, String hierarchyName) {...}

public static ILevel getLevel(IActivePivot pivot, String levelDescription) {...}

public static ILevel getLevel(final IActivePivot pivot, final String dimensionName, final String hierarchyName, final String levelName) {...}

Filters on analysis hierarchies, PostProcessors and V2

In 5.7 the behavior of aggregate retrievers has been changed: the retrievals with filters on an analysis hierarchy now return null. In 5.7.3 we introduced new versions of post processors (ABasicPostProcessorV2, ABaseDynamicAggregationPostProcessorV2, ADynamicAggregationPostProcessorV2 and AFilteringPostProcessorV2) that automatically remove the filters on analysis hierarchies in the prefetcher of the post processor, and add the filter back while expanding the result on the analysis hierarchies. It was introduced as V2 to avoid changing the behavior of the post processor in a bugfix release.

In 5.8 the post processors V2 have been removed and the core post processors ABasicPostProcessor, ABaseDynamicAggregationPostProcessor, ADynamicAggregationPostProcessor and AFilteringPostProcessor now perform this filter removal if the "analysisLevels" property is used.

Keep in mind that AAdvancedPostProcessor does not handle those filters. If you implement AAdvancedPostProcessor directly, this confluence page explains how to handle them.

In order to solve this problem, ABasicPostProcessor (previously ABaseDynamicPostProcessors) now uses named prefetchers. They will use the constant ABasicPostProcessor.BASIC_POST_PROCESSOR_PREFETCHER ( previously ABaseDynamicAggregationPostProcessor.DYNAMIC_AGGREGATION_POST_PROCESSOR_PREFETCHER) as the name of the prefetcher to retrieve. If you define your own prefetcher you need to name it correctly:

IPrefetcher<?> namedPrefetcher = IPrefetcher.name(ABasicPostProcessor.BASIC_POST_PROCESSOR_PREFETCHER, prefetcher);

CubeFilterBuilder

ICubeFilterBuilder.includeMembers([...]) and excludeMembers([...]) now only accept List<String> as members path name argument. The previous signature that accepted List<?> as members has been moved to ICubeFilterBuilder.includeMembersWithConditions([...]) and excludeMembersWithConditions([...]).

The includeMembersWithConditions([...]) and excludeMembersWithConditions([...]) methods should only be used when building the CubeFilter with IConditions.

Range Sharing configuration

Range sharing can no longer be configured with a boolean. You must use an integer or use the withRangeSharing() method on the cube builder, that is, .withProperty("rangeSharing", "false") has been replaced by .withoutRangeSharing().

AStoreStream

Listeners on a continuous query no longer receive an initial view when registering to this query. This was not the case with com.quartetfs.biz.pivot.postprocessing.streams.impl.AStoreStream, and has been fixed. AStoreStream is a selection listener that registers on a datastore selection. This operation can occur in the middle of a transaction. The javadoc of com.quartetfs.biz.pivot.postprocessing.streams.impl.AStoreStream#registerSelection gives some pointers for concrete implementations.

Datastore

Duplicate key handling

Replacement: IDuplicateKeyWithinTransactionListenerIDuplicateKeyHandler

The IDuplicateKeyWithinTransactionListener interface has been replaced by the IDuplicateKeyHandler interface, as described in the Changelog.

With IDuplicateKeyWithinTransactionListener you could only define a custom behavior when two records had the same key fields within the transaction but no record with these key fields existed in the datastore yet. You can still define such a behavior using the IDuplicateKeyHandler.selectDuplicateKeyWithinTransaction function. However, it is now also possible to define a custom behavior when a record in the transaction has the same key fields as a record already in the datastore, by defining the IDuplicateKeyHandler.selectDuplicateKeyInDatastore function.

The default behavior, which is to always update with the latest record, has not changed.

The store description builder has been modified: you should replace .onDuplicateKeyWithinTransaction().logException() with .withDuplicateKeyHandler(DuplicateKeyHandlers.LOG_WITHIN_TRANSACTION)

Selection Builder

When using the datastore selection builder, you can add all fields reachable by references from a given store using .withAllreachableFields(). The only policy used to be that all fields with same name had to be reachable from one another, and that the furthest would be used. The selection builder now walks the field graph using references between stores, and keeps track of the paths leading to the same field. If there are no collisions, nothing happens. If there are, three options are available:

  • resolve conflicts automatically by choosing the furthest field and throw an exception if two fields with the same name are not reachable from one another (example: ref1/field and ref2/field). This is how it used to be in 5.7 and is the default behavior.
  • resolve conflicts automatically by choosing the closest field and throw an exception if two fields with the same name are not reachable from one another (example: ref1/field and ref2/field).
  • resolve conflicts manually. You can use a FieldsCollisionHandler to transform a map of {field name, [possible paths]} into a map of {name, full expression} (the expression is of the form ref1/ref2/fieldName). This allows you to keep one, the other, both (with custom field name), or none.

Test jar

The datastore test utils that were previously packaged in a test-jar have been moved to their own module:

 <groupId>com.activeviam.tech</groupId>
<artifactId>datastore-test</artifactId>

Removal of LINE storage

The in-line storage type provided by ActivePivot 5.0 as a safe alternative to the new columnar storage has been removed. Only one value remains possible: StorageType.COLUMN.

Changes are required only if your Datastore stores were explicitly configured to use the LINE storage. If so, either replace it with COLUMN as below, or remove the call to configure the storage to use the default version.

new StoreDescriptionBuilder().withStoreName("<store>")
.withField(...).asKeyField()
... // Other configuration calls
- .withStorage(IStoreDescription.StorageType.LINE)
+ .withStorage(IStoreDescription.StorageType.COLUMN)
...
.build();

Format of partitions sizes

StoreUtils.getPartitionSizes(storeVersion) now returns values of -1 in the array for non-existing partitions (either dropped or never initialized) at the given storeVersion.

JDBCSource

The constructors for IJDBCSource implementations without an appendBatchSize have been removed.

The constructors for IJDBCSource implementations no longer require a driverClassName in the arguments when an IConnectionSupplier is provided.

The SimpleConnectionSupplier implementation of IConnectionSupplier now requires a driverClassName argument.

The appendQueueSize attribute has been removed from all IJDBCSource implementation constructors, and is now a field of the JDBCTopic class.

JSON Web Tokens (JWT)

The RSA key pair defined by the qfs.jwt.key.[public|private] variables in jwt.properties should now be encoded with at least 2048 bits instead of the previous 1024. A new pair can be generated using JwtUtil.

StreamingService

The signature of IStreamingService.updateStreamRanges changed. It now takes an IAxisRanges instead of an array of IRangeProperties.

Spring configuration

With the introduction of the centralized property accessor ActiveViamProperty, the environment properties will be available only if you import the ActiveViamPropertyFromSpringConfig configuration class. For more information, see Properties in ActivePivot.

Memory analysis

With 5.8.0, Memory Analysis services have been improved. Some elements have been migrated to the dedicated application Memory Analysis Cube - DatastoreFeederVisitor, other classes have been renamed to better emphasize that the services are designed for Memory analysis only and not monitoring.

Renamed classes:

  • IMemoryMonitoringService -> IMemoryAnalysisService
  • MemoryMonitoringService -> MemoryAnalysisService
  • MonitoringStatisticSerializerUtil -> MemoryStatisticSerializerUtil

Removed classes:

For the sake of readability, only major classes are mentioned.

  • DatastoreFeederVisitor
  • MemoryMonitoringDatastoreDescription

Previous migration notes

You can find the previous Migration Notes in our old Confluence documentation: