The Wayback Machine - https://web.archive.org/web/20200424023914/https://github.com/topics/scala
Skip to content
#

Scala

scala logo

Scala is a general-purpose programming language providing support for functional programming and a strong static type system. Designed to be concise, many of Scala's design decisions aimed to address criticisms of Java.

Here are 9,736 public repositories matching this topic...

erkankarabulut
erkankarabulut commented Apr 15, 2020

I am trying to deploy the app with the given ./sbt clean dist command but I got this error:

Downloading sbt launcher for 1.3.8:
  From  https://repo.scala-sbt.org/scalasbt/maven-releases/org/scala-sbt/sbt-launch/1.3.8/sbt-launch-1.3.8.jar
    To  /root/.sbt/launchers/1.3.8/sbt-launch.jar
Downloading sbt launcher 1.3.8 md5 hash:
  From  https://repo.scala-sbt.org/scalasbt/maven-releas
SIkebe
SIkebe commented Jan 2, 2020

Issue

Impacted version: 4.33.0

Deployment mode: standalone app

Repro steps:

  1. Wiki -> New Page
  2. Fill in each item and click "Save" button.
    • Page name: title+
    • Content: aaa
  3. Redirect to the edit page.
    • Page name: title instead of title+
    • Content: (blank)

As + is decoded to whitespace, should it be treated as an invalid character he

bug
adriancole
adriancole commented Aug 29, 2018

Let's support at least reading "b3" header from a single string, most commonly traceid-spanid-1
It would also be nice to support optionally writing this, especially in message providers or others with constrained environments.

Expected behavior

As discussed on openzipkin/b3-propagation#21 and first implemented here: https://github.com/openzipkin/brave/blob/master/brave/src/main/java/bra

lk-geimfari
lk-geimfari commented Aug 30, 2019

Well, we need to check all the projects and remove outdated ones.

We declare this in the readme:

However, keep in mind that we don't accept mammoth's shit. Only active and interesting projects with good documentation are added. Dead and abandoned projects will be removed.

But, sadly at this moment, this project looks like a mammoth shit itself.

We really need to fix it. I think

lila
fennuzhichui
fennuzhichui commented Jan 2, 2020

how to define java8 when submit application use spark-submit


name: Bug report/Feature request/Question
about: Create a report to help us improve
title: ''
label: bug/enhancement/question
assignees: ''

Environment:

  • Java version:
  • Scala version:
  • Spark version:
  • PyTorch and Python version:
  • OS and version:

Checklist:

  • Did you check if your bug/feature/
NeQuissimus
NeQuissimus commented Jan 7, 2019

I just came across the fact that a Show[Throwable] exists.

  1. It is not wired up into import Scalaz._, one needs to import scalaz.std.java.throwable._
  2. It discards the stack trace entirely.

What's going on with this? :D
Is it OK if I fix both of these (for 7.2 and 7.3)?

LeeTibbert
LeeTibbert commented Apr 6, 2020

Once Issue #1749 is merged, build.sbt should be edited to add the incantation
Global / onChangedBuildSource := ReloadOnSourceChanges

This is available as of sbt 1.3.8. I find it reduces my cycle time & frustration
when making edits to build.sbt, etc. Basically, I do not have to go through a
5 or 15 minute failure cycle because I forgot to do a manual reload. With that
setting, sbt det

yiheng
yiheng commented Jul 11, 2018

Spark 2.3 officially support run on kubernetes. While our guide of "Run on Kubernetes" is still based on a special version of Spark 2.2, which is out of date. We need to:

  1. update that document to Spark 2.3
  2. release the corresponding docker images.
casperdcl
casperdcl commented Jul 16, 2019

perhaps it would be nice to provide an example where injected parameters are inside a dictionary or object:

# parameters
import argparse
args = argparse.Namespace()
args_dict = {}

args.a = 1
args_dict['b'] = 2
papermill ... -p args.a 1.618 -p "args_dict['b']" 3.14159

This would of course be useful for making transition between *.py scripts (using e.g

yeikel
yeikel commented Jan 4, 2019

According to the generated build

The commands to launch are the following :

docker pull andypetrella/spark-notebook:0.7.0-scala-2.11.8-spark-2.1.1-hadoop-2.7.2-with-hive
docker run -p 9001:9001 andypetrella/spark-notebook:0.7.0-scala-2.11.8-spark-2.1.1-hadoop-2.7.2-with-hive

Using that image (and I think it i

Created by Martin Odersky

Released January 20, 2004

Website
www.scala-lang.org
Wikipedia
Wikipedia
You can’t perform that action at this time.