Merge pull request #1215 from eikek/dep-updates

Dep updates
This commit is contained in:
mergify[bot] 2021-12-11 22:00:26 +00:00 committed by GitHub
commit 05843876cc
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
30 changed files with 46 additions and 45 deletions

View File

@ -1,4 +1,4 @@
version = "3.0.6" version = "3.0.8"
preset = default preset = default
align.preset = some align.preset = some

View File

@ -14,7 +14,7 @@ def inTest(d0: Seq[ModuleID], ds: Seq[ModuleID]*) =
val scalafixSettings = Seq( val scalafixSettings = Seq(
semanticdbEnabled := true, // enable SemanticDB semanticdbEnabled := true, // enable SemanticDB
semanticdbVersion := scalafixSemanticdb.revision, //"4.4.0" semanticdbVersion := scalafixSemanticdb.revision, // "4.4.0"
ThisBuild / scalafixDependencies ++= Dependencies.organizeImports ThisBuild / scalafixDependencies ++= Dependencies.organizeImports
) )
@ -781,7 +781,7 @@ val restserver = project
Universal / mappings := { Universal / mappings := {
val allMappings = (Universal / mappings).value val allMappings = (Universal / mappings).value
allMappings.filter { allMappings.filter {
//scalajs artifacts are not needed at runtime // scalajs artifacts are not needed at runtime
case (file, name) => !name.contains("_sjs1_") case (file, name) => !name.contains("_sjs1_")
} }
} }

View File

@ -53,8 +53,8 @@ object Properties {
"ner.statisticalOnly" -> "true", "ner.statisticalOnly" -> "true",
"ner.rulesOnly" -> "false", "ner.rulesOnly" -> "false",
"ner.applyFineGrained" -> "false", "ner.applyFineGrained" -> "false",
"ner.applyNumericClassifiers" -> "false", //only english supported, not needed currently "ner.applyNumericClassifiers" -> "false", // only english supported, not needed currently
"ner.useSUTime" -> "false", //only english, unused in docspell "ner.useSUTime" -> "false", // only english, unused in docspell
"ner.language" -> "de", "ner.language" -> "de",
"ner.model" -> "edu/stanford/nlp/models/ner/german.distsim.crf.ser.gz,edu/stanford/nlp/models/ner/english.conll.4class.distsim.crf.ser.gz" "ner.model" -> "edu/stanford/nlp/models/ner/german.distsim.crf.ser.gz,edu/stanford/nlp/models/ner/english.conll.4class.distsim.crf.ser.gz"
).withRegexNer(regexNerMappingFile).withHighRecall(highRecall) ).withRegexNer(regexNerMappingFile).withHighRecall(highRecall)

View File

@ -50,7 +50,7 @@ object JobState {
NonEmptyList.of(Waiting, Scheduled, Running, Stuck, Failed, Cancelled, Success) NonEmptyList.of(Waiting, Scheduled, Running, Stuck, Failed, Cancelled, Success)
val queued: Set[JobState] = Set(Waiting, Scheduled, Stuck) val queued: Set[JobState] = Set(Waiting, Scheduled, Stuck)
val done: NonEmptyList[JobState] = NonEmptyList.of(Failed, Cancelled, Success) val done: NonEmptyList[JobState] = NonEmptyList.of(Failed, Cancelled, Success)
val notDone: NonEmptyList[JobState] = //all - done val notDone: NonEmptyList[JobState] = // all - done
NonEmptyList.of(Waiting, Scheduled, Running, Stuck) NonEmptyList.of(Waiting, Scheduled, Running, Stuck)
val inProgress: Set[JobState] = Set(Scheduled, Running, Stuck) val inProgress: Set[JobState] = Set(Scheduled, Running, Stuck)

View File

@ -168,7 +168,7 @@ object MimeType {
left left
} }
//https://datatracker.ietf.org/doc/html/rfc7230#section-3.2.6 // https://datatracker.ietf.org/doc/html/rfc7230#section-3.2.6
private def isToken(s: String): Boolean = private def isToken(s: String): Boolean =
s.nonEmpty && s.forall(c => c.isLetterOrDigit || tokenExtraChars.contains(c)) s.nonEmpty && s.forall(c => c.isLetterOrDigit || tokenExtraChars.contains(c))
@ -186,7 +186,7 @@ object MimeType {
seq(primary, sub)((p, s) => MimeType(p.toLowerCase, s.toLowerCase, None)) seq(primary, sub)((p, s) => MimeType(p.toLowerCase, s.toLowerCase, None))
} }
//https://datatracker.ietf.org/doc/html/rfc2046#section-4.1.2 // https://datatracker.ietf.org/doc/html/rfc2046#section-4.1.2
private val charset: P[Option[Charset]] = in => private val charset: P[Option[Charset]] = in =>
in.trim.toLowerCase.indexOf("charset=") match { in.trim.toLowerCase.indexOf("charset=") match {
case -1 => Right((None, in)) case -1 => Right((None, in))

View File

@ -44,7 +44,7 @@ object PdfExtract {
s"Using stripped text (not OCR), as it is longer (${strippedRes._1.length} > ${ocrStr.length})" s"Using stripped text (not OCR), as it is longer (${strippedRes._1.length} > ${ocrStr.length})"
) *> Result(strippedRes).pure[F] ) *> Result(strippedRes).pure[F]
//maybe better: inspect the pdf and decide whether ocr or not // maybe better: inspect the pdf and decide whether ocr or not
for { for {
pdfboxRes <- pdfboxRes <-
logger.debug("Trying to strip text from pdf using pdfbox.") *> logger.debug("Trying to strip text from pdf using pdfbox.") *>

View File

@ -16,8 +16,8 @@ import munit._
class ImageSizeTest extends FunSuite { class ImageSizeTest extends FunSuite {
//tiff files are not supported on the jdk by default // tiff files are not supported on the jdk by default
//requires an external library // requires an external library
val files = List( val files = List(
ExampleFiles.camera_letter_en_jpg -> Dimension(1695, 2378), ExampleFiles.camera_letter_en_jpg -> Dimension(1695, 2378),
ExampleFiles.camera_letter_en_png -> Dimension(1695, 2378), ExampleFiles.camera_letter_en_png -> Dimension(1695, 2378),

View File

@ -15,8 +15,8 @@ object Playing extends IOApp {
def run(args: List[String]): IO[ExitCode] = def run(args: List[String]): IO[ExitCode] =
IO { IO {
//val ods = ExampleFiles.examples_sample_ods.readURL[IO](8192, blocker) // val ods = ExampleFiles.examples_sample_ods.readURL[IO](8192, blocker)
//val odt = ExampleFiles.examples_sample_odt.readURL[IO](8192, blocker) // val odt = ExampleFiles.examples_sample_odt.readURL[IO](8192, blocker)
val rtf = ExampleFiles.examples_sample_rtf.readURL[IO](8192) val rtf = ExampleFiles.examples_sample_rtf.readURL[IO](8192)
val x = for { val x = for {

View File

@ -46,7 +46,7 @@ object RegexNerFile {
final private class Impl[F[_]: Async]( final private class Impl[F[_]: Async](
cfg: Config, cfg: Config,
store: Store[F], store: Store[F],
writer: Semaphore[F] //TODO allow parallelism per collective writer: Semaphore[F] // TODO allow parallelism per collective
) extends RegexNerFile[F] { ) extends RegexNerFile[F] {
def makeFile(collective: Ident): F[Option[Path]] = def makeFile(collective: Ident): F[Option[Path]] =

View File

@ -143,7 +143,7 @@ object CreateItem {
.fromOption[F](NonEmptyList.fromList(fileMetaIds.toList)) .fromOption[F](NonEmptyList.fromList(fileMetaIds.toList))
.flatMap(fids => .flatMap(fids =>
OptionT( OptionT(
//load attachments but only those mentioned in the task's arguments // load attachments but only those mentioned in the task's arguments
cand.headOption.traverse(ri => cand.headOption.traverse(ri =>
ctx.store ctx.store
.transact(RAttachment.findByItemCollectiveSource(ri.id, ri.cid, fids)) .transact(RAttachment.findByItemCollectiveSource(ri.id, ri.cid, fids))
@ -195,7 +195,7 @@ object CreateItem {
ctx.logger.error(msg) *> Sync[F].raiseError(new Exception(msg)) ctx.logger.error(msg) *> Sync[F].raiseError(new Exception(msg))
} }
//TODO if no source is present, it must be saved! // TODO if no source is present, it must be saved!
private def originFileTuple( private def originFileTuple(
t: (RAttachment, Option[RAttachmentSource]) t: (RAttachment, Option[RAttachmentSource])
): (Ident, Ident) = ): (Ident, Ident) =

View File

@ -60,7 +60,7 @@ object EvalProposals {
): Double = ): Double =
mp.proposalType match { mp.proposalType match {
case MetaProposalType.DueDate => case MetaProposalType.DueDate =>
//for due dates, sort earliest on top // for due dates, sort earliest on top
MetaProposal MetaProposal
.parseDate(cand) .parseDate(cand)
.map { ld => .map { ld =>

View File

@ -108,9 +108,9 @@ object ReProcessItem {
data.item.cid, data.item.cid,
args.itemId.some, args.itemId.some,
lang, lang,
None, //direction None, // direction
data.item.source, //source-id data.item.source, // source-id
None, //folder None, // folder
Seq.empty, Seq.empty,
false, false,
None, None,

View File

@ -162,7 +162,7 @@ object ScanMailboxTask {
def requireFolder[C](a: Access[F, C])(name: String): MailOp[F, C, MailFolder] = def requireFolder[C](a: Access[F, C])(name: String): MailOp[F, C, MailFolder] =
if ("INBOX".equalsIgnoreCase(name)) a.getInbox if ("INBOX".equalsIgnoreCase(name)) a.getInbox
else //TODO resolve sub-folders else // TODO resolve sub-folders
a.findFolder(None, name) a.findFolder(None, name)
.map(_.toRight(new Exception(s"Folder '$name' not found"))) .map(_.toRight(new Exception(s"Folder '$name' not found")))
.mapF(_.rethrow) .mapF(_.rethrow)

View File

@ -227,7 +227,7 @@ final class SchedulerImpl[F[_]: Async](
job.id, job.id,
config.name, config.name,
store store
) //also increments retries if current state=stuck ) // also increments retries if current state=stuck
def wrapTask( def wrapTask(
job: RJob, job: RJob,

View File

@ -26,7 +26,7 @@ class NerFileTest extends FunSuite {
for ((name, first) <- names) { for ((name, first) <- names) {
val ps = Pattern(1)(name).distinct val ps = Pattern(1)(name).distinct
//check if it compiles to a regex pattern // check if it compiles to a regex pattern
ps.flatMap(_.value.split("\\s+").toList).foreach(_.r) ps.flatMap(_.value.split("\\s+").toList).foreach(_.r)
ps.foreach(_.value.r) ps.foreach(_.value.r)

View File

@ -59,7 +59,7 @@ class JsonMiniQueryTest extends FunSuite with Fixtures {
assertEquals(r(sampleEvent), Vector.empty) assertEquals(r(sampleEvent), Vector.empty)
} }
//content.[added,removed].(category=expense & name=grocery) // content.[added,removed].(category=expense & name=grocery)
test("combine fields and filter") { test("combine fields and filter") {
val andOk = JQ.at("content").at("added", "removed") >> val andOk = JQ.at("content").at("added", "removed") >>
(JQ.at("name").is("grocery") && JQ.at("category").is("expense")) (JQ.at("name").is("grocery") && JQ.at("category").is("expense"))
@ -156,6 +156,6 @@ class JsonMiniQueryTest extends FunSuite with Fixtures {
val json4 = parseJson( val json4 = parseJson(
"""[{"name":"max", "count":4}, {"name":"me", "count": 3}, {"name":"max", "count": 3}]""" """[{"name":"max", "count":4}, {"name":"me", "count": 3}, {"name":"max", "count": 3}]"""
) )
println(q4(json4)) assertEquals(q4(json4), values("max", "max"))
} }
} }

View File

@ -80,7 +80,7 @@ final class NaivePubSub[F[_]: Async](
} yield head } yield head
def publish(topic: Topic): Pipe[F, Json, MessageHead] = def publish(topic: Topic): Pipe[F, Json, MessageHead] =
ms => //TODO Do some optimization by grouping messages to the same topic ms => // TODO Do some optimization by grouping messages to the same topic
ms.evalMap(publish0(topic, _)) ms.evalMap(publish0(topic, _))
def subscribe(topics: NonEmptyList[Topic]): Stream[F, Message[Json]] = def subscribe(topics: NonEmptyList[Topic]): Stream[F, Message[Json]] =

View File

@ -58,7 +58,7 @@ class NaivePubSubTest extends CatsEffectSuite with Fixtures {
res <- subscribe(ps, Topics.jobSubmitted) res <- subscribe(ps, Topics.jobSubmitted)
(received, halt, subFiber) = res (received, halt, subFiber) = res
_ <- ps.publish1(otherTopic, JobSubmittedMsg("hello".id)) _ <- ps.publish1(otherTopic, JobSubmittedMsg("hello".id))
_ <- IO.sleep(100.millis) //allow some time for receiving _ <- IO.sleep(100.millis) // allow some time for receiving
_ <- halt.set(true) _ <- halt.set(true)
outcome <- subFiber.join outcome <- subFiber.join
_ = assert(outcome.isSuccess) _ = assert(outcome.isSuccess)

View File

@ -113,7 +113,7 @@ object ExprString {
case Expr.NamesMacro(name) => case Expr.NamesMacro(name) =>
s"${C.names}:${quote(name)}" s"${C.names}:${quote(name)}"
case Expr.YearMacro(_, year) => case Expr.YearMacro(_, year) =>
s"${C.year}:$year" //currently, only for Attr.Date s"${C.year}:$year" // currently, only for Attr.Date
case Expr.ConcMacro(term) => case Expr.ConcMacro(term) =>
s"${C.conc}:${quote(term)}" s"${C.conc}:${quote(term)}"
case Expr.CorrMacro(term) => case Expr.CorrMacro(term) =>

View File

@ -236,9 +236,9 @@ trait Conversions {
i.concPerson.map(mkIdName), i.concPerson.map(mkIdName),
i.concEquip.map(mkIdName), i.concEquip.map(mkIdName),
i.folder.map(mkIdName), i.folder.map(mkIdName),
Nil, //attachments Nil, // attachments
Nil, //tags Nil, // tags
Nil, //customfields Nil, // customfields
i.notes, i.notes,
Nil // highlight Nil // highlight
) )

View File

@ -141,7 +141,7 @@ object ContentDisposition {
} }
private val parser = makeParser(mimeValue) private val parser = makeParser(mimeValue)
//private val origParser = makeParser(Rfc7230.token | Rfc7230.quotedString) // private val origParser = makeParser(Rfc7230.token | Rfc7230.quotedString)
implicit val headerInstance: Header[ContentDisposition, Header.Single] = { implicit val headerInstance: Header[ContentDisposition, Header.Single] = {
val oh = `Content-Disposition`.headerInstance val oh = `Content-Disposition`.headerInstance

View File

@ -140,9 +140,9 @@ object IntegrationEndpointRoutes {
cfg: Config.IntegrationEndpoint.AllowedIps cfg: Config.IntegrationEndpoint.AllowedIps
): HttpRoutes[F] = ): HttpRoutes[F] =
HttpRoutes { req => HttpRoutes { req =>
//The `req.from' take the X-Forwarded-For header into account, // The `req.from' take the X-Forwarded-For header into account,
//which is not desirable here. The `http-header' auth config // which is not desirable here. The `http-header' auth config
//can be used to authenticate based on headers. // can be used to authenticate based on headers.
val from = req.remote.map(_.host) val from = req.remote.map(_.host)
if (from.exists(cfg.containsAddress)) OptionT.none[F, Response[F]] if (from.exists(cfg.containsAddress)) OptionT.none[F, Response[F]]
else OptionT.pure(Responses.forbidden[F]) else OptionT.pure(Responses.forbidden[F])

View File

@ -81,7 +81,7 @@ trait MigrationTasks {
def mkTransactor(ctx: Context): Transactor[IO] = { def mkTransactor(ctx: Context): Transactor[IO] = {
val xa = Transactor.fromConnection[IO](ctx.getConnection()) val xa = Transactor.fromConnection[IO](ctx.getConnection())
Transactor.strategy.set(xa, Strategy.void) //transactions are handled by flyway Transactor.strategy.set(xa, Strategy.void) // transactions are handled by flyway
} }
} }

View File

@ -61,7 +61,7 @@ object QJob {
retryPause: Duration, retryPause: Duration,
currentTry: Int currentTry: Int
): F[Either[Unit, Option[RJob]]] = { ): F[Either[Unit, Option[RJob]]] = {
//if this fails, we have to restart takeNextJob // if this fails, we have to restart takeNextJob
def markJob(job: RJob): F[Either[Unit, RJob]] = def markJob(job: RJob): F[Either[Unit, RJob]] =
store.transact(for { store.transact(for {
n <- RJob.setScheduled(job.id, worker) n <- RJob.setScheduled(job.id, worker)
@ -236,7 +236,7 @@ object QJob {
val JC = RJob.T val JC = RJob.T
val waiting = NonEmptyList.of(JobState.Waiting, JobState.Stuck, JobState.Scheduled) val waiting = NonEmptyList.of(JobState.Waiting, JobState.Stuck, JobState.Scheduled)
val running = NonEmptyList.of(JobState.Running) val running = NonEmptyList.of(JobState.Running)
//val done = JobState.all.filterNot(js => ).diff(waiting).diff(running) // val done = JobState.all.filterNot(js => ).diff(waiting).diff(running)
def selectJobs(now: Timestamp): Stream[ConnectionIO, RJob] = { def selectJobs(now: Timestamp): Stream[ConnectionIO, RJob] = {
val refDate = now.minusHours(24) val refDate = now.minusHours(24)

View File

@ -20,7 +20,7 @@ import doobie.implicits._
* a 0..1-1 relationship. * a 0..1-1 relationship.
*/ */
case class RAttachmentArchive( case class RAttachmentArchive(
id: Ident, //same as RAttachment.id id: Ident, // same as RAttachment.id
fileId: Ident, fileId: Ident,
name: Option[String], name: Option[String],
messageId: Option[String], messageId: Option[String],

View File

@ -17,7 +17,7 @@ import doobie._
import doobie.implicits._ import doobie.implicits._
case class RAttachmentMeta( case class RAttachmentMeta(
id: Ident, //same as RAttachment.id id: Ident, // same as RAttachment.id
content: Option[String], content: Option[String],
nerlabels: List[NerLabel], nerlabels: List[NerLabel],
proposals: MetaProposalList, proposals: MetaProposalList,

View File

@ -19,7 +19,7 @@ import doobie.implicits._
* 1-1 (or 0..1-1) relationship. * 1-1 (or 0..1-1) relationship.
*/ */
case class RAttachmentPreview( case class RAttachmentPreview(
id: Ident, //same as RAttachment.id id: Ident, // same as RAttachment.id
fileId: Ident, fileId: Ident,
name: Option[String], name: Option[String],
created: Timestamp created: Timestamp

View File

@ -19,7 +19,7 @@ import doobie.implicits._
* 1-1 (or 0..1-1) relationship. * 1-1 (or 0..1-1) relationship.
*/ */
case class RAttachmentSource( case class RAttachmentSource(
id: Ident, //same as RAttachment.id id: Ident, // same as RAttachment.id
fileId: Ident, fileId: Ident,
name: Option[String], name: Option[String],
created: Timestamp created: Timestamp

View File

@ -120,7 +120,7 @@ object RCollective {
cs.listType.s, cs.listType.s,
es.schedule.s, es.schedule.s,
es.minAge.s, es.minAge.s,
const(0) //dummy value to load Nil as list of passwords const(0) // dummy value to load Nil as list of passwords
), ),
from(c).leftJoin(cs, cs.cid === c.id).leftJoin(es, es.cid === c.id), from(c).leftJoin(cs, cs.cid === c.id).leftJoin(es, es.cid === c.id),
c.id === coll c.id === coll

View File

@ -15,7 +15,7 @@ object Dependencies {
val DoobieVersion = "1.0.0-RC1" val DoobieVersion = "1.0.0-RC1"
val EmilVersion = "0.10.0-M3" val EmilVersion = "0.10.0-M3"
val FlexmarkVersion = "0.62.2" val FlexmarkVersion = "0.62.2"
val FlywayVersion = "8.2.0" val FlywayVersion = "8.2.1"
val Fs2Version = "3.2.3" val Fs2Version = "3.2.3"
val Fs2CronVersion = "0.7.1" val Fs2CronVersion = "0.7.1"
val H2Version = "1.4.200" val H2Version = "1.4.200"
@ -288,7 +288,8 @@ object Dependencies {
// https://github.com/flyway/flyway // https://github.com/flyway/flyway
// ASL 2.0 // ASL 2.0
val flyway = Seq( val flyway = Seq(
"org.flywaydb" % "flyway-core" % FlywayVersion "org.flywaydb" % "flyway-core" % FlywayVersion,
"org.flywaydb" % "flyway-mysql" % FlywayVersion
) )
val yamusca = Seq( val yamusca = Seq(