big-data
Here are 2,331 public repositories matching this topic...
-
Updated
Jan 29, 2021
-
Updated
Jan 28, 2021 - Python
-
Updated
Jan 31, 2021 - JavaScript
-
Updated
Jan 9, 2021 - Scala
-
Updated
Dec 16, 2020 - Scala
The latest copy of the CPython grammar tests in test_grammar.py has several @skips and FIXMEs. Some of them seem easy to fix, e.g. some parser bugs or missing warnings that would be helpful, others are entire features. We should fix the easy ones and make sure there are tickets for the rest.
Problem: the approximate method can still be slow for many trees
catboost version: master
Operating System: ubuntu 18.04
CPU: i9
GPU: RTX2080
Would be good to be able to specify how many trees to use for shapley. The model.predict and prediction_type versions allow this. lgbm/xgb allow this.
-
Updated
Jan 31, 2021 - Jupyter Notebook
-
Updated
Jan 31, 2021 - Go
-
Updated
Jan 29, 2021 - Erlang
- Remove wasted white space on either side of table
- Support adding 1 or 2 more sub fields
- Allow hiding columns
- Resize columns
-
Updated
Sep 1, 2020 - Python
Please describe the problem you are trying to solve
I would like to evict entries based on their creation time. I want to evict the oldest ones first.
Please describe the desired behavior
Basically FIFO eviction. I would like to specify directly in the configuration something like:
<eviction eviction-policy="FIFO" max-size-policy="PER_NODE" size="5000"/>
**Describe alte
-
Updated
Jan 21, 2021 - Scala
-
Updated
Jan 31, 2021 - Java
PrestoDB https://prestodb.io .. is widely used as SQL frontend for many different data-sources, including ElasticSearch, and even files in S3 .. would be very nice if there would be a Connector available for Vespa.
Hi, if my spark app is using 2 storage type, both S3 and Azure Data Lake Store Gen2, could I put spark.delta.logStore.class=org.apache.spark.sql.delta.storage.AzureLogStore, org.apache.spark.sql.delta.storage.S3SingleDriverLogStore
Thanks in advance
-
Updated
Jan 30, 2021 - TypeScript
-
Updated
Jan 12, 2021 - JavaScript
Currently, we test Parquet with default config, but not with the optimized writer.
See #6382.
Improve this page
Add a description, image, and links to the big-data topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the big-data topic, visit your repo's landing page and select "manage topics."


Now insert and query share the resource ( Max Process Count control) 。 When the query with high TPS,the insert will get error (“error: too many process”). I think separator the resource for Insert and Query will makes sense. Ensure enough resource for insert。It looks like Use Yarn, Insert and Query use the different resource quota。
Or the simple way , Can we set Ratio for Insert and