-
Updated
Dec 3, 2021 - Jupyter Notebook
mlops
Here are 517 public repositories matching this topic...
-
Updated
Dec 4, 2021
-
Updated
Nov 19, 2021
-
Updated
Dec 16, 2021 - Jupyter Notebook
Describe the bug
data docs columns shrink to 1 character width with long query
To Reproduce
Steps to reproduce the behavior:
- make a batch from a long query string
- run validation
- render result to data docs
- See screenshot
<img width="1525" alt="Data_documentation_compiled_by_Great_Expectations" src="https://user-images.githubusercontent.com/928247/103230647-30eca500-4
With a config like this
{
"METAFLOW_DATASTORE_SYSROOT_S3": "s3://mf-test/metaflow/",
}
(note a slash after METAFLOW_DATASTORE_SYSROOT_S3)
metaflow.S3(run=self).put* produces double-slashes like here:
s3://mf-test/metaflow//data/DataLoader/1630978962283843/month=01/data.parquet
The trailing slash in the config shouldn't make a difference
Description
We're running usability tests and would love for you to record walking through our tutorials. The idea for this ticket is that you do a screen capture walking through one of more of the following examples:
- Hello World! (15 minutes)
- [Iris Dataset](https://kedro.readthedocs.io/en/stable/02_get_started
🚨 🚨 Feature Request
- Related to an existing Issue
- A new implementation (Improvement, Extension)
If your feature will improve HUB
Need a way to check if a dataset already exists.
hub.empty throws an error if a dataset exists and hub.load throws an error if the dataset does not exist.
Need a way to check if a dataset already exists without throwing a
-
Updated
Dec 16, 2021 - Python
-
Updated
Dec 17, 2021 - Python
For SC Operator it may be a good idea to generate CRD manifests from inside a docker container.
This should provide reproducible generation step and avoid "produces different output on my machine" issues.
Linter should also fail if generation of manifests produce diff with the commited version.
Is your feature request related to a problem? Please describe.
I have a framework that handles the offline store. It creates the tables, indexes, reads data from different data sources, does some transformations, and then inserts into the offline store. As a part of this, I can construct the entities, feature views, feature services, etc, a instance of the ParsedRepo class for Feast. What I n
Translate the Chinese deployment doc into the English version. The English version should be put into docs/en/deploy.md .
-
Updated
Dec 17, 2021 - Go
Need to make utilities in aim/sdk/num_utils.py to treat as numeric values the following types:
numpy.ndarraywith shape(1,).- subclasses of
numpy.number. - tensor for scalar values (
tensorflow,torch).
Change Run.track() method to not allow values which are not numeric values nor AimObject.
Describe the issue
Currently we run the Linter CI for golang repos using the golang-ci-linter binary. But according to the documentation it is faster and better to use the github action.
https://golangci-lint.run/usage/install/
The UX of all the error highlighting is also better.
What if we do not do this?
Finding linter errors is troublesome as users have to parse through the cons
-
Updated
Dec 17, 2021 - Jupyter Notebook
-
Updated
Dec 17, 2021 - Python
-
Updated
Dec 1, 2021 - Kotlin
-
Updated
Aug 23, 2021 - Python
Currently, there is no way to limit the number of tokens using the config file.
Potential config could contain JSON element like:
"dataset": {
"columns": [
{
"name": "id",
"type": "number"
},
{
"name": "author",
"type": "text",
---------> "tokenLimit": 1000, <-----------------
},
{
"name"
-
Updated
Dec 6, 2021 - Python
-
Updated
Nov 30, 2021 - Jupyter Notebook
-
Updated
Dec 16, 2021 - Python
Users can tell Ploomber to track changes to configuration files via resources_, however, to track changes, we compute a file hash which may take too long if the file is large.
We should show a warning if this happens, resources_ should not be used with large files.
Improve this page
Add a description, image, and links to the mlops topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the mlops topic, visit your repo's landing page and select "manage topics."


At the moment, from API there are two useful columns about tasks:
However, in tabs (views) there are possible columns
tasks:created_ata