We invite you to shape the future of IBM, including product roadmaps, by submitting ideas that matter to you the most. Here's how it works:
Post your ideas
Post ideas and requests to enhance a product or service. Take a look at ideas others have posted and upvote them if they matter to you,
Post an idea
Upvote ideas that matter most to you
Get feedback from the IBM team to refine your idea
Help IBM prioritize your ideas and requests
The IBM team may need your help to refine the ideas so they may ask for more information or feedback. The product management team will then decide if they can begin working on your idea. If they can start during the next development cycle, they will put the idea on the priority list. Each team at IBM works on a different schedule, where some ideas can be implemented right away, others may be placed on a different schedule.
Receive notification on the decision
Some ideas can be implemented at IBM, while others may not fit within the development plans for the product. In either case, the team will let you know as soon as possible. In some cases, we may be able to find alternatives for ideas which cannot be implemented in a reasonable time.
Python packages can currently be added using the approach described in https://www.ibm.com/docs/en/cloud-paks/cp-data/3.5.0?topic=packages-customizing-using-user-home-volume However, some Python packages use shared libraries that may not be instal...
Spark job will be fail when resource quota not enough
There are lots of Spark Job to run, if a job start and the resource quota is reach maximum, this job will fail. You will wait for job result to know is quota is not enough or you need to check quota every time when you submit a job. For AlibabaClo...
The Spark job should use the mount_path attribute of the storage volume by default
The volumes.mount_path attribute in the Spark Jobs API is mandatory, https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=jobs-spark-api-syntax-parameters-return-codes This should be an optional attribute to make it consistent with the usage o...
Extend policy and rule enforcement to Analytics Engine
Analytics Engine components such as Hive are an important part of the Data Lake story. Without Hive, external users are unable to query data in IBM Cloud Object Storage.
However, Analytics Engine is not integrated with Watson Knowledge Catalog. T...
Graphframes package for Spark provides the capabilities of both the GraphX and the Spark DataFrames. The recent Apache Spark 3.0 release has several enhancements with respect to the features enablement and performance, especially, for the Spark Da...
HDInsight make it super easy to use Hive ACID transactions - can we make it easier for users to setup hive transactions on Analytics Engine?
Hive LLAP provides large performance improvements. Can we add to the IAE documentation how to configure it?
Azure make this especially easy by provide LLAP by default with certain cluster types - https://docs.microsoft.com/en-us/azure/hdinsight/i...
Do not place IBM confidential, company confidential, or personal information into any field.