We invite you to shape the future of IBM, including product roadmaps, by submitting ideas that matter to you the most. Here's how it works:
Post your ideas
Post ideas and requests to enhance a product or service. Take a look at ideas others have posted and upvote them if they matter to you,
Post an idea
Upvote ideas that matter most to you
Get feedback from the IBM team to refine your idea
Help IBM prioritize your ideas and requests
The IBM team may need your help to refine the ideas so they may ask for more information or feedback. The product management team will then decide if they can begin working on your idea. If they can start during the next development cycle, they will put the idea on the priority list. Each team at IBM works on a different schedule, where some ideas can be implemented right away, others may be placed on a different schedule.
Receive notification on the decision
Some ideas can be implemented at IBM, while others may not fit within the development plans for the product. In either case, the team will let you know as soon as possible. In some cases, we may be able to find alternatives for ideas which cannot be implemented in a reasonable time.
In the provided ticket (TS005406891) we are looking for the ability to configure the Nvidia Compute Mode via a Spectrum Conductor setting/config. This would make managing the setting simpler via a config change versus requiring each customer to im...
We are in immediate need of latest pyarrow version for ppc64le and we are looking for v2.0.Also we need a conda package ( NOT PIP) and will appreciate if this could be made available to us at the earliest. Having said this would like to follow up ...
Add slot demand as part of the spark application metric pushed out to elastic search
We would like to have slot demand included as part of the spark application metric that are currently pushed out to elastic search which would help us identify the slot demand vs provisioned for the spark job submitted by the users.
Requesting support for application level performance metrics for Dask jobs to be loaded into ELK. Currently, we get these metrics for spark jobs and would like to get this type of information for Dask jobs as well in 2.5.
We would like to use pytest framework to run test cases that contains python notebooks and essentially some of the notebooks use spark and h2o. We would like to understand if you are aware how to make this work using conductor.
Conductor 2.5 log retrieval feature back porting to Conductor 2.4.1. That is, to allow specifying arbitrary limit for number of executors to lookup in the RESTful call to retrieve specific executor log by its executor's id only.
Need for a better, configurable dashboard, that communicates the (usable) slots and memory utilization of the grid, provides some context, shows where the actual work is happening, and can be used by developers, grid admins, and regular business u...
Do not place IBM confidential, company confidential, or personal information into any field.