This portal is to open public enhancement requests against products and services offered by the IBM Data Platform organization. To view all of your ideas submitted to IBM, create and manage groups of Ideas, or create an idea explicitly set to be either visible by all (public) or visible only to you and IBM (private), use the IBM Unified Ideas Portal (https://ideas.ibm.com).
Shape the future of IBM!
We invite you to shape the future of IBM, including product roadmaps, by submitting ideas that matter to you the most. Here's how it works:
Search existing ideas
Start by searching and reviewing ideas and requests to enhance a product or service. Take a look at ideas others have posted, and add a comment, vote, or subscribe to updates on them if they matter to you. If you can't find what you are looking for,
Post your ideas
Post ideas and requests to enhance a product or service. Take a look at ideas others have posted and upvote them if they matter to you,
Post an idea
Upvote ideas that matter most to you
Get feedback from the IBM team to refine your idea
Specific links you will want to bookmark for future use
Welcome to the IBM Ideas Portal (https://www.ibm.com/ideas) - Use this site to find out additional information and details about the IBM Ideas process and statuses.
IBM Unified Ideas Portal (https://ideas.ibm.com) - Use this site to view all of your ideas, create new ideas for any IBM product, or search for ideas across all of IBM.
ideasibm@us.ibm.com - Use this email to suggest enhancements to the Ideas process or request help from IBM for submitting your Ideas.
IBM Employees should enter Ideas at https://ideas.ibm.com
See this idea on ideas.ibm.com
In a customer case after only 9 month the file-api-claim had over 800k files and it was suspected that we had a performance issue.
We disabled selinux relabling
But we also ran the cron job cleanup script.
But this was not enough. We had to run more cleanup manually to remove spark jobs.
We went from 800k files to 500k
It would be nice that we add this functionality into the product to auto-clean.
Also scripts reffered here does did not go very deep in the file cleaning with CP4D 4.8.4, it could be better enhanced after we discussed with Manjot from the spark dev team.
https://wwwpoc.ibm.com/support/pages/node/6980928
Maybe we could put in the gui for the admin a usability option where the customer sets how many days of logs he wants to keep.
IBM would automatically clean up for the customer by default in the product, maybe we could keep 2 days as stated in the script.
Needed By | Not sure -- Just thought it was cool |
By clicking the "Post Comment" or "Submit Idea" button, you are agreeing to the IBM Ideas Portal Terms of Use.
Do not place IBM confidential, company confidential, or personal information into any field.
Other tip on asset-file-api cleanup
https://www.ibm.com/support/pages/large-amount-files-stored-filesystem-causes-performance-degradation-and-some-cases-cluster-unavailability
Spark has added feature in 5.1 to enhance this
https://www.ibm.com/docs/en/cloud-paks/cp-data/5.1.x?topic=applications-configuring-spark-log-level-information
Also see tip https://wwwpoc.ibm.com/support/pages/node/6980928
This is a step in the right direction, we need other services to participate into this request.
Would be nice if we can set how long we want to retain logs in the CP4D / Software hub GUI and CP4D would handle clean up daily or weekly.
2 more issues related to this
TS018772895 and TS018664310
This is a high priority feature request.
Log and zip files (from project exports even using --hard-delete) has always been a problem (since CPD3.2) causing the file-api-claim to grow without bounds FOR NO REASON. This is very bad behavior for any serious application.