data-engineering
Here are 1,021 public repositories matching this topic...
-
Updated
Nov 30, 2021
-
Updated
Aug 14, 2021
-
Updated
May 28, 2021
Description
It is not an actual bug but in the documentation here -> https://docs.prefect.io/orchestration/concepts/api.html#queries
flow_run actually needs to be flow_runs.
Otherwise it does not work for me.
Expected Behavior
Documentation should be updated.
Reproduction
Describe the bug
data docs columns shrink to 1 character width with long query
To Reproduce
Steps to reproduce the behavior:
- make a batch from a long query string
- run validation
- render result to data docs
- See screenshot
<img width="1525" alt="Data_documentation_compiled_by_Great_Expectations" src="https://user-images.githubusercontent.com/928247/103230647-30eca500-4
Tell us about the problem you're trying to solve
We can probably reduce the docker image size of our java based connectors by using the ADD command instead of COPYing the tar archive. See this PR for an example
Describe the solution you’d like
use the ADD command to reduce the size of the docker images
-
Updated
Nov 29, 2021 - Go
When specifying on demand feature views at retrieval time (e.g. get_X_features), the output feature vectors include e.g. request data or dependent feature vectors, even if users did not specify said features.
Expected Behavior
Non-specified dependent feature values are not returned in output
Current Behavior
Non-specified dependent feature values are in output
Steps to reprodu
-
Updated
Nov 29, 2021 - Python
-
Updated
Oct 29, 2021
What
being able to take a data object (or prefix, like a partition) and get back the commit that added/modified it.
Why
This is valuable lineage information that is currently available in lakeFS but not exposed easily, and mimics the behavior of git blame
How
Given the lakeFS API already supports listing the log of commits for an object or prefix (
-
Updated
Aug 2, 2021 - JavaScript
-
Updated
Nov 30, 2021 - Jupyter Notebook
-
Updated
Nov 15, 2021
-
Updated
Nov 29, 2021 - Jupyter Notebook
-
Updated
Mar 9, 2020 - Python
if they are not class methods then the method would be invoked for every test and a session would be created for each of those tests.
`class PySparkTest(unittest.TestCase):
@classmethod
def suppress_py4j_logging(cls):
logger = logging.getLogger('py4j')
logger.setLevel(logging.WARN)
@classmethod
def create_testing_pyspark_session(cls):
return Sp
Hi ,
I am using some basic functions from pyjanitor such as - clean_names() , collapse_levels() in one of my code which I want to productionise.
And there are limitations on the size of the production code base.
Currently ,if I just look at the requirements.txt for just "pyjanitor" , its huge .
I don't think I require all the dependencies in my code.
How can I remove the unnecessary ones ?
-
Updated
Nov 27, 2021
The load_dotted_path raises the following error if unable to load the module:
Traceback (most recent call last):
File "/Users/Edu/Desktop/import-error/script.py", line 4, in <module>
load_dotted_path('tests.quality.fn')
File "/Users/Edu/dev/ploomber/src/ploomber/util/dotted_path.py", line 128, in load_dotted_path
module = importlib.import_module(mod)
File "/Users/-
Updated
Jun 2, 2021
-
Updated
Mar 5, 2020 - Python
-
Updated
Oct 25, 2021
-
Updated
Nov 29, 2021 - Python
-
Updated
Nov 6, 2021 - Ruby
Is there an existing issue for this?
- I have searched the existing issues
Current Behavior
A large amount of output goes to the log, this should not happen by default.
Expected Behavior
much less content in the output of the FVT and the build bu default
Switch on debug in the logging configuration and then see all the output.
Steps To Reproduce
run the build
Env
-
Updated
Nov 22, 2021 - TypeScript
-
Updated
Feb 7, 2021 - CSS
Improve this page
Add a description, image, and links to the data-engineering topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the data-engineering topic, visit your repo's landing page and select "manage topics."



A clear and concise description of what the bug is.
The superset chart table sets the number of pagination rows to select the setting, and add the option of whether to select all
like this
![image](https://user-images.githubusercontent.com/52438024/141405677-f9e25aef-e0d3-4e99-986a