Skip to end of metadata
Go to start of metadata
You are viewing an old version of this page. View the current version.
Compare with Current
View Page History
« Previous
Version 22
Next »
Datameer X 11.2.12
Improvements
1 | Add Support for Cloudera Private Cloud / Cloudera Runtime 7.1.9.0 | DAP-42564 |
---|
| Datameer now supports Cloudera Private Cloud / Cloudera Runtime 7.1.9.0 |
|
2 | Add support for Amazon AWS EMR 6.10.0 | DAP-42587 |
---|
| Datameer now supports Amazon AWS EMR version 6.10.0. |
|
3 | Add support for Amazon AWS EMR 6.11.0 | DAP-42586 |
---|
| Datameer now supports Amazon AWS EMR version 6.11.0. |
|
4 | Add support for Amazon AWS EMR 6.12.0 | DAP-42585 |
---|
| Datameer now supports Amazon AWS EMR version 6.12.0. |
|
5 | Add support for Amazon AWS EMR 6.13.0 | DAP-42584 |
---|
| Datameer now supports Amazon AWS EMR version 6.13.0. |
|
5 | Add support for Amazon AWS EMR 6.15.0 | DAP-42581 |
---|
| Datameer now supports Amazon AWS EMR version 6.15.0. |
|
Bug Fixes
1 | Joins on BigDecimal keys return different results when Map side vs Reduce side join is used | DAP-42574 |
---|
| BigDecimal values of different scales are considered equal for JOIN and GROUPBY operations. | Bug fixed. |
Datameer X 11.2.11
Improvements
1 | Supported Hadoop Distributions: Add support for Amazon AWS EMR 6.14.0 | DAP-42564 |
---|
| Datameer now supports Amazon AWS EMR version 6.14.0. |
|
2 | Security: Upgrade JSch Library for SFPT/SCP connections | DAP-42550 |
---|
| The Java SSH implementation has been updated to fulfill the latest security recommendations. |
|
Bug Fixes
1 | Joins on BigDecimal keys return different results when Map side vs Reduce side join is used | DAP-42574 |
---|
| BigDecimal values of different scales are considered equal for JOIN and GROUPBY operations. | Bug fixed. |
Datameer X 11.2.10
Improvements
Datameer X 11.2.9
Bug Fixes
Datameer X 11.2.8
Improvements
Datameer X 11.2.7
Improvements
1 | Plug-ins: plugin-snowflake - use Snowflake's internal stage for exports | DAP-42491 |
---|
| The plugin now provides the option to select Snowflake's internal stage as the default storage option. After writing the export data into the local file system, it will be imported to Snowflake via a 'PUT' command. |
|
Bug Fixes
1 | Timezone not resolving correctly in Mexico | DAP-42497 |
---|
| After fixing code, the correct timezone is now displayed again for date and time assets in Mexico. | Bug fixed. |
2 | Workbook: Sorting by the 'Last Processed' time doesn't work | DAP-42501 |
---|
| The sorting functionality now works again as expected in the File Browser's Artifacts bar. | Bug fixed. |
Datameer X 11.2.6
Improvements
1 | REST API: Output a response when creating a Workbook vie the API | DAP-42481 |
---|
| REST API v2 calls for creating a new Workbook now return the following information: status, configuration ID, file ID and file UUID, and the. file path. |
|
2 | Import/ Export: Use the local Datameer timezone instead of UTC while converting numeric values into a date | DAP-42475 |
---|
| Two new global properties ensure that Parquet int96/int64 codes timestamps are treated as being in Datameer's timezone now when set to 'false': 'das.import.parquet.int64.timestamp.adjusted-to-utc' and 'das.import.parquet.int96.timestamp.adjusted-to-utc'. |
|
3 | Supported Hadoop Distributions: Add support for Cloudera Private Cloud/ Cloudera Runtime 7.1.8 | DAP-42394 |
---|
| Datameer now supports Cloudera Private Cloud/ Cloudera Runtime 7.1.8. |
|
Bug Fixes
1 | Administration: Housekeeping doesn't work - jobs stuck in the job scheduler | DAP-42485 & DAP-42489 |
---|
| Housekeeping can now be configured via the properties 'housekeeping.keep-deleted-data' and 'housekeeping.keep-deleted-data-max' to remove outdated files. | Bug fixed. |
2 | REST API: Workbooks with schedule created via REST API v2 are not triggered | DAP-42484 |
---|
| Scheduled triggering for Workbooks that are created via the REST API v2 is now working without a need to manually resave the Workbook's configuration in the User Interface. | Bug fixed. |
3 | Import/ Export: Updating Snowflake drivers | DAP-42479 |
---|
| The patched plugin 'snowflake-jdbc' is now working to enable the Datameer - Snowflake connection again. | Bug fixed. |
Datameer X 11.2.5
Improvements
1 | Security: Upgrade Jetty to version 9.4.50 | DAP-42458 |
---|
| Datameer is moved to the Jetty version 9.4.50 in order to ensure the services are not vulnerable to any known
vulnerabilities. |
|
Bug Fixes
1 | Security: Penetration tests - GetFiles metadata, get the folder content via the folder ID | DAP-42468 |
---|
| A user is not allowed to view any metadata for a directory without proper access (sees no folder path or file ID). | Bug fixed. |
2 | Security: Penetration tests - Any user may not see others schema log | DAP-42469 |
---|
| A user is not allowed to see the parsing records from the artifacts created by other users. | Bug fixed. |
3 | Security: Penetration tests - The '/browser/list-file' endpoint returns metadata for any file in the system | DAP-42470 |
---|
| A user is not allowed to view artifacts metadata without having at least a 'View' permission for these items. | Bug fixed. |
4 | Security: Penetration tests - Dependency graph information doesn't check the permission authorization | DAP-42471 |
---|
| A user is not allowed to load an artifact’s dependency graph without having at least a 'View' permission for this item. | Bug fixed. |
5 | Security: Penetration tests - detailed error messages displayed | DAP-42472 |
---|
| Only generic error messages without error details are now returned to Datameer users. Stack traces are logged server-side and only accessible by developers or administrators. The associated property handles the error messages. | Bug fixed. |
6 | Security: Penetration tests - '/rest/user-management/authenticable-users' should return list of usernames belonging to the same role only with CHANGE_FOLDER_OWNER capability | DAP-42473 |
---|
| Only user who belong to the same role with the 'change folder owner' capability can view other user names from the REST API call. | Bug fixed. |
Datameer X 11.2.4
Improvements
1 | Drivers: Update Redshift Native JDBC driver to use Amazon's JDBC42 driver | DAP-42443 |
---|
| Since the Amazon Redshift Native JDCB driver is deprecated. Therefore the the new drive should be updated to the Amazon Redshift JDBC42 driver. |
|
Bug Fixes
1 | Plugins: plugin-hbase+cdh-7.1.7.0 - Import fails | DAP-42440 |
---|
| Setting the property 'hbase.server.sasl.provider.extras=org.apache.hadoop.hbase.security.provider.GssSaslClientAuthenticationProvider' bypasses the service loader lookup code with providing authentication providers. | Bug fixed. |
Datameer X 11.2.3
Improvements
1 | Importing Data: Add an option to configure the default value for "Records for Schema Detection" | DAP-42421 |
---|
| The new custom property 'das.conductor.default.record.sample.size' controls what 'Records for schema detection' value is set by default whenever a user creates a new Import Job, Data Link or File Upload. It should be adjusted, e.g. when working with large JSON objects (should be decreased to 250 - 500). Once updated, it affects only newly created Import Jobs, the 'Records for schema detection' value for existing artifacts remains intact. |
|
Bug Fixes
1 | Export: Snowflake - Exporting to a new table leads to duplicated columns error intermittently | DAP-42012 |
---|
| After fixing the plug-in, the export sheet can be executed without any errors again. | Bug fixed. |
2 | Export: Snowflake - Intermittent Export failure, insert value list does not match the column list | DAP-42442 |
---|
| After a plug-in fix, the columns list now matches the insert value list as expected again. | Bug fixed. |
Datameer X 11.2.2
Improvements
1 | Supported Hadoop Distributions: Add support for Amazon EMR 6.8.0 | DAP-42400 |
---|
| Datameer X now supports Amazon EMR version 6.7.0. |
|
2 | Properties: Make HadoopUtil#HADOOP_ACCESS_TIMEOUT configurable | DAP-42404 |
---|
| The timeout in a S3A filesystem can be changed via the properties 'fs.s3a.connection.establish.timout' and 'fs.s3a.connection.timeout' now in order to prevent job timeouts. |
|
Bug Fixes
1 | Security: Spring framework CVE vulnerabilities | DAP-42402 |
---|
| The user has been backported to Spring framework 4.3. | Bug fixed. |
Datameer X 11.2.1
Bug Fixes
1 | Import/ Export: Hive (with ADLS backed external table) - Failure on import | DAP-42389 |
---|
| Hive import jobs now succeeds again when executed at the cluster after setting several properties and adapted Hadoop configuration. | Bug fixed. |
2 | Plug-ins: Plugin Parquet - Datameer encodes DATE values in base64 while exporting into a Parquet file | DAP-42382 |
---|
| After the 'plugin-parquet' was patched, the DATE fields are exported correctly. | Bug fixed. |
3 | Export: Tableau - Hyper jobs fail intermittently with "Hyper Server did not call back on the callback connection." | DAP-42380 |
---|
| Users are now able to export hyper formatted files to Tableau without any failures again after the associated plug-in has been updated. | Bug fixed. |
Datameer X 11.2.0
Improvements
1 | Supported Hadoop Distributions: Add support for Amazon EMR 5.36.0 | DAP-42369 |
---|
| Datameer X now supports Amazon EMR version 5.36.0. |
|
2 | Supported Hadoop Distributions: Add support for Amazon EMR 6.7.0 | DAP-42375 |
---|
| Datameer X now supports Amazon EMR version 6.7.0. |
|
3 | Backup & Restore: Performance and lack of logging problem | DAP-42370 |
---|
| Performance issues during the backup and restore process are now executed at an expected speed for large number of artifacts. |
|
Bug Fixes
1 | Plug-ins: EMR - 'plugin-emr' High Availability discovery mode picks local running Yarn Ressource Manager at 0.0.0.0:8088 | DAP-42378 |
---|
| The EMR cluster can now be set up in High Availability mode. | Bug fixed. |
2 | Plug-ins: EMR - 'plugin-emr' "default=obtain from ec2 environment" region option cannot be saved | DAP-42372 |
---|
| When configuring the cluster mode in EMR, the default option for the region can now be saved. | Bug fixed. |
3 | Upgrade: Java upgrade 'Upgrade Filesystem Artifact To Delete' doesn't trigger the database upgrade | DAP-42 |
---|
| The Java upgrade script now updates the table 'Upgrade Filesystem Artifact To Delete' as expected or throws a clear error message. | Bug fixed. |