Experiment with addons

Addons allow to execute external programs in some context inside
docspell. Currently it is possible to run them after processing files.
Addons are provided by URLs to zip files.
This commit is contained in:
eikek 2022-04-22 14:07:28 +02:00
parent e04a76faa4
commit 7fdd78ad06
166 changed files with 8181 additions and 115 deletions

View File

@ -293,6 +293,15 @@ val openapiScalaSettings = Seq(
field.copy(typeDef =
TypeDef("DownloadState", Imports("docspell.common.DownloadState"))
)
case "addon-trigger-type" =>
field =>
field.copy(typeDef =
TypeDef("AddonTriggerType", Imports("docspell.addons.AddonTriggerType"))
)
case "addon-runner-type" =>
field =>
field
.copy(typeDef = TypeDef("RunnerType", Imports("docspell.addons.RunnerType")))
})
)
@ -325,6 +334,7 @@ val common = project
libraryDependencies ++=
Dependencies.fs2 ++
Dependencies.circe ++
Dependencies.circeGenericExtra ++
Dependencies.calevCore ++
Dependencies.calevCirce
)
@ -351,7 +361,7 @@ val files = project
.in(file("modules/files"))
.disablePlugins(RevolverPlugin)
.settings(sharedSettings)
.withTestSettings
.withTestSettingsDependsOn(loggingScribe)
.settings(
name := "docspell-files",
libraryDependencies ++=
@ -448,6 +458,19 @@ val notificationApi = project
)
.dependsOn(common, loggingScribe)
val addonlib = project
.in(file("modules/addonlib"))
.disablePlugins(RevolverPlugin)
.settings(sharedSettings)
.withTestSettingsDependsOn(loggingScribe)
.settings(
libraryDependencies ++=
Dependencies.fs2 ++
Dependencies.circe ++
Dependencies.circeYaml
)
.dependsOn(common, files, loggingScribe)
val store = project
.in(file("modules/store"))
.disablePlugins(RevolverPlugin)
@ -469,7 +492,16 @@ val store = project
libraryDependencies ++=
Dependencies.testContainer.map(_ % Test)
)
.dependsOn(common, query.jvm, totp, files, notificationApi, jsonminiq, loggingScribe)
.dependsOn(
common,
addonlib,
query.jvm,
totp,
files,
notificationApi,
jsonminiq,
loggingScribe
)
val notificationImpl = project
.in(file("modules/notification/impl"))
@ -647,7 +679,7 @@ val restapi = project
openapiSpec := (Compile / resourceDirectory).value / "docspell-openapi.yml",
openapiStaticGen := OpenApiDocGenerator.Redoc
)
.dependsOn(common, query.jvm, notificationApi, jsonminiq)
.dependsOn(common, query.jvm, notificationApi, jsonminiq, addonlib)
val joexapi = project
.in(file("modules/joexapi"))
@ -667,7 +699,7 @@ val joexapi = project
openapiSpec := (Compile / resourceDirectory).value / "joex-openapi.yml",
openapiStaticGen := OpenApiDocGenerator.Redoc
)
.dependsOn(common, loggingScribe)
.dependsOn(common, loggingScribe, addonlib)
val backend = project
.in(file("modules/backend"))
@ -683,6 +715,7 @@ val backend = project
Dependencies.emil
)
.dependsOn(
addonlib,
store,
notificationApi,
joexapi,
@ -739,7 +772,7 @@ val config = project
Dependencies.fs2 ++
Dependencies.pureconfig
)
.dependsOn(common, loggingApi, ftspsql, store)
.dependsOn(common, loggingApi, ftspsql, store, addonlib)
// --- Application(s)
@ -946,6 +979,7 @@ val root = project
)
.aggregate(
common,
addonlib,
loggingApi,
loggingScribe,
config,

View File

@ -0,0 +1,90 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.addons
import cats.effect._
import cats.syntax.all._
import fs2.Stream
import fs2.io.file.{Files, Path}
import docspell.common._
import docspell.files.Zip
final case class AddonArchive(url: LenientUri, name: String, version: String) {
def nameAndVersion: String =
s"$name-$version"
def extractTo[F[_]: Async](
reader: UrlReader[F],
directory: Path,
withSubdir: Boolean = true,
glob: Glob = Glob.all
): F[Path] = {
val logger = docspell.logging.getLogger[F]
val target =
if (withSubdir) directory.absolute / nameAndVersion
else directory.absolute
Files[F]
.exists(target)
.flatMap {
case true => target.pure[F]
case false =>
Files[F].createDirectories(target) *>
reader(url)
.through(Zip.unzip(8192, glob))
.through(Zip.saveTo(logger, target, moveUp = true))
.compile
.drain
.as(target)
}
}
/** Read meta either from the given directory or extract the url to find the metadata
* file to read
*/
def readMeta[F[_]: Async](
urlReader: UrlReader[F],
directory: Option[Path] = None
): F[AddonMeta] =
directory
.map(AddonMeta.findInDirectory[F])
.getOrElse(AddonMeta.findInZip(urlReader(url)))
}
object AddonArchive {
def read[F[_]: Async](
url: LenientUri,
urlReader: UrlReader[F],
extractDir: Option[Path] = None
): F[AddonArchive] = {
val addon = AddonArchive(url, "", "")
addon
.readMeta(urlReader, extractDir)
.map(m => addon.copy(name = m.meta.name, version = m.meta.version))
}
def dockerAndFlakeExists[F[_]: Async](
archive: Either[Path, Stream[F, Byte]]
): F[(Boolean, Boolean)] = {
val files = Files[F]
def forPath(path: Path): F[(Boolean, Boolean)] =
(files.exists(path / "Dockerfile"), files.exists(path / "flake.nix")).tupled
def forZip(data: Stream[F, Byte]): F[(Boolean, Boolean)] =
data
.through(Zip.unzip(8192, Glob("Dockerfile|flake.nix")))
.collect {
case bin if bin.name == "Dockerfile" => (true, false)
case bin if bin.name == "flake.nix" => (false, true)
}
.compile
.fold((false, false))((r, e) => (r._1 || e._1, r._2 || e._2))
archive.fold(forPath, forZip)
}
}

View File

@ -0,0 +1,33 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.addons
import cats.Monoid
import cats.syntax.all._
case class AddonExecutionResult(
addonResults: List[AddonResult],
pure: Boolean
) {
def addonResult: AddonResult = addonResults.combineAll
def isFailure: Boolean = addonResult.isFailure
def isSuccess: Boolean = addonResult.isSuccess
}
object AddonExecutionResult {
val empty: AddonExecutionResult =
AddonExecutionResult(Nil, false)
def combine(a: AddonExecutionResult, b: AddonExecutionResult): AddonExecutionResult =
AddonExecutionResult(
a.addonResults ::: b.addonResults,
a.pure && b.pure
)
implicit val executionResultMonoid: Monoid[AddonExecutionResult] =
Monoid.instance(empty, combine)
}

View File

@ -0,0 +1,121 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.addons
import cats.data.Kleisli
import cats.effect._
import cats.syntax.all._
import fs2.Stream
import fs2.io.file._
import docspell.common.UrlReader
import docspell.common.exec.Env
import docspell.logging.Logger
trait AddonExecutor[F[_]] {
def config: AddonExecutorConfig
def execute(logger: Logger[F]): AddonExec[F]
def execute(logger: Logger[F], in: InputEnv): F[AddonExecutionResult] =
execute(logger).run(in)
}
object AddonExecutor {
def apply[F[_]: Async](
cfg: AddonExecutorConfig,
urlReader: UrlReader[F]
): AddonExecutor[F] =
new AddonExecutor[F] with AddonLoggerExtension {
val config = cfg
def execute(logger: Logger[F]): AddonExec[F] =
Kleisli { in =>
for {
_ <- logger.info(s"About to run ${in.addons.size} addon(s) in ${in.baseDir}")
ctx <- prepareDirectory(
logger,
in.baseDir,
in.outputDir,
in.cacheDir,
in.addons
)
rs <- ctx.traverse(c => runAddon(logger.withAddon(c), in.env)(c))
pure = ctx.foldl(true)((b, c) => b && c.meta.isPure)
} yield AddonExecutionResult(rs, pure)
}
private def prepareDirectory(
logger: Logger[F],
baseDir: Path,
outDir: Path,
cacheDir: Path,
addons: List[AddonRef]
): F[List[Context]] =
for {
addonsDir <- Directory.create(baseDir / "addons")
_ <- Directory.createAll(Context.tempDir(baseDir), outDir, cacheDir)
_ <- Context
.userInputFile(baseDir)
.parent
.fold(().pure[F])(Files[F].createDirectories)
archives = addons.map(_.archive).distinctBy(_.url)
_ <- logger.info(s"Extract ${archives.size} addons to $addonsDir")
mkCtxs <- archives.traverse { archive =>
for {
_ <- logger.debug(s"Extracting $archive")
addonDir <- archive.extractTo(urlReader, addonsDir)
meta <- AddonMeta.findInDirectory(addonDir)
mkCtx = (ref: AddonRef) =>
Context(ref, meta, baseDir, addonDir, outDir, cacheDir)
} yield archive.url -> mkCtx
}
ctxFactory = mkCtxs.toMap
res = addons.map(ref => ctxFactory(ref.archive.url)(ref))
} yield res
private def runAddon(logger: Logger[F], env: Env)(
ctx: Context
): F[AddonResult] =
for {
_ <- logger.info(s"Executing addon ${ctx.meta.nameAndVersion}")
_ <- logger.trace("Storing user input into file")
_ <- Stream
.emit(ctx.addon.args)
.through(fs2.text.utf8.encode)
.through(Files[F].writeAll(ctx.userInputFile, Flags.Write))
.compile
.drain
runner <- selectRunner(cfg, ctx.meta, ctx.addonDir)
result <- runner.run(logger, env, ctx)
} yield result
}
def selectRunner[F[_]: Async](
cfg: AddonExecutorConfig,
meta: AddonMeta,
addonDir: Path
): F[AddonRunner[F]] =
for {
addonRunner <- meta.enabledTypes(Left(addonDir))
// intersect on list retains order in first
possibleRunner = cfg.runner
.intersect(addonRunner)
.map(AddonRunner.forType[F](cfg))
runner = possibleRunner match {
case Nil =>
AddonRunner.failWith(
s"No runner available for addon config ${meta.runner} and config ${cfg.runner}."
)
case list =>
AddonRunner.firstSuccessful(list)
}
} yield runner
}

View File

@ -0,0 +1,45 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.addons
import docspell.addons.AddonExecutorConfig._
import docspell.common.Duration
import docspell.common.exec.{Args, SysCmd}
case class AddonExecutorConfig(
runner: List[RunnerType],
runTimeout: Duration,
nspawn: NSpawn,
nixRunner: NixConfig,
dockerRunner: DockerConfig
)
object AddonExecutorConfig {
case class NSpawn(
enabled: Boolean,
sudoBinary: String,
nspawnBinary: String,
containerWait: Duration
) {
val nspawnVersion =
SysCmd(nspawnBinary, Args.of("--version")).withTimeout(Duration.seconds(2))
}
case class NixConfig(
nixBinary: String,
buildTimeout: Duration
)
case class DockerConfig(
dockerBinary: String,
buildTimeout: Duration
) {
def dockerBuild(imageName: String): SysCmd =
SysCmd(dockerBinary, "build", "-t", imageName, ".").withTimeout(buildTimeout)
}
}

View File

@ -0,0 +1,28 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.addons
import docspell.logging.Logger
trait AddonLoggerExtension {
implicit final class LoggerAddonOps[F[_]](self: Logger[F]) {
private val addonName = "addon-name"
private val addonVersion = "addon-version"
def withAddon(r: AddonArchive): Logger[F] =
self.capture(addonName, r.name).capture(addonVersion, r.version)
def withAddon(r: Context): Logger[F] =
withAddon(r.addon.archive)
def withAddon(m: AddonMeta): Logger[F] =
self.capture(addonName, m.meta.name).capture(addonVersion, m.meta.version)
}
}
object AddonLoggerExtension extends AddonLoggerExtension

View File

@ -0,0 +1,216 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.addons
import java.io.FileNotFoundException
import cats.data.OptionT
import cats.effect._
import cats.syntax.all._
import fs2.Stream
import fs2.io.file.{Files, Path}
import docspell.common.Glob
import docspell.files.Zip
import io.circe.generic.semiauto.{deriveDecoder, deriveEncoder}
import io.circe.yaml.{parser => YamlParser}
import io.circe.{Decoder, Encoder}
import io.circe.{parser => JsonParser}
case class AddonMeta(
meta: AddonMeta.Meta,
triggers: Option[Set[AddonTriggerType]],
args: Option[List[String]],
runner: Option[AddonMeta.Runner],
options: Option[AddonMeta.Options]
) {
def nameAndVersion: String =
s"${meta.name}-${meta.version}"
def parseResult: Boolean =
options.exists(_.collectOutput)
def ignoreResult: Boolean =
!parseResult
def isImpure: Boolean =
options.exists(_.isImpure)
def isPure: Boolean =
options.forall(_.isPure)
/** Returns a list of runner types that are possible to use for this addon. This is also
* inspecting the archive to return defaults when the addon isn't declaring it in the
* descriptor.
*/
def enabledTypes[F[_]: Async](
archive: Either[Path, Stream[F, Byte]]
): F[List[RunnerType]] =
for {
filesExists <- AddonArchive.dockerAndFlakeExists(archive)
(dockerFileExists, flakeFileExists) = filesExists
nixEnabled = runner.flatMap(_.nix).map(_.enable) match {
case Some(flag) => flag
case None => flakeFileExists
}
dockerEnabled = runner.flatMap(_.docker).map(_.enable) match {
case Some(flag) => flag
case None => dockerFileExists
}
trivialEnabled = runner.flatMap(_.trivial).exists(_.enable)
result = RunnerType.all.filter(_.fold(nixEnabled, dockerEnabled, trivialEnabled))
} yield result
}
object AddonMeta {
def empty(name: String, version: String): AddonMeta =
AddonMeta(Meta(name, version, None), None, None, None, None)
case class Meta(name: String, version: String, description: Option[String])
case class Runner(
nix: Option[NixRunner],
docker: Option[DockerRunner],
trivial: Option[TrivialRunner]
)
case class NixRunner(enable: Boolean)
case class DockerRunner(enable: Boolean, image: Option[String], build: Option[String])
case class TrivialRunner(enable: Boolean, exec: String)
case class Options(networking: Boolean, collectOutput: Boolean) {
def isPure = !networking && collectOutput
def isImpure = networking
def isUseless = !networking && !collectOutput
def isUseful = networking || collectOutput
}
object NixRunner {
implicit val jsonEncoder: Encoder[NixRunner] =
deriveEncoder
implicit val jsonDecoder: Decoder[NixRunner] =
deriveDecoder
}
object DockerRunner {
implicit val jsonEncoder: Encoder[DockerRunner] =
deriveEncoder
implicit val jsonDecoder: Decoder[DockerRunner] =
deriveDecoder
}
object TrivialRunner {
implicit val jsonEncoder: Encoder[TrivialRunner] =
deriveEncoder
implicit val jsonDecoder: Decoder[TrivialRunner] =
deriveDecoder
}
object Runner {
implicit val jsonEncoder: Encoder[Runner] =
deriveEncoder
implicit val jsonDecoder: Decoder[Runner] =
deriveDecoder
}
object Options {
implicit val jsonEncoder: Encoder[Options] =
deriveEncoder
implicit val jsonDecoder: Decoder[Options] =
deriveDecoder
}
object Meta {
implicit val jsonEncoder: Encoder[Meta] =
deriveEncoder
implicit val jsonDecoder: Decoder[Meta] =
deriveDecoder
}
implicit val jsonEncoder: Encoder[AddonMeta] =
deriveEncoder
implicit val jsonDecoder: Decoder[AddonMeta] =
deriveDecoder
def fromJsonString(str: String): Either[Throwable, AddonMeta] =
JsonParser.decode[AddonMeta](str)
def fromJsonBytes[F[_]: Sync](bytes: Stream[F, Byte]): F[AddonMeta] =
bytes
.through(fs2.text.utf8.decode)
.compile
.string
.map(fromJsonString)
.rethrow
def fromYamlString(str: String): Either[Throwable, AddonMeta] =
YamlParser.parse(str).flatMap(_.as[AddonMeta])
def fromYamlBytes[F[_]: Sync](bytes: Stream[F, Byte]): F[AddonMeta] =
bytes
.through(fs2.text.utf8.decode)
.compile
.string
.map(fromYamlString)
.rethrow
def findInDirectory[F[_]: Sync: Files](dir: Path): F[AddonMeta] = {
val logger = docspell.logging.getLogger[F]
val jsonFile = dir / "docspell-addon.json"
val yamlFile = dir / "docspell-addon.yaml"
val yamlFile2 = dir / "docspell-addon.yml"
OptionT
.liftF(Files[F].exists(jsonFile))
.flatTap(OptionT.whenF(_)(logger.debug(s"Reading json addon file $jsonFile")))
.flatMap(OptionT.whenF(_)(fromJsonBytes(Files[F].readAll(jsonFile))))
.orElse(
OptionT
.liftF(Files[F].exists(yamlFile))
.flatTap(OptionT.whenF(_)(logger.debug(s"Reading yaml addon file $yamlFile")))
.flatMap(OptionT.whenF(_)(fromYamlBytes(Files[F].readAll(yamlFile))))
)
.orElse(
OptionT
.liftF(Files[F].exists(yamlFile2))
.flatTap(OptionT.whenF(_)(logger.debug(s"Reading yaml addon file $yamlFile2")))
.flatMap(OptionT.whenF(_)(fromYamlBytes(Files[F].readAll(yamlFile2))))
)
.getOrElseF(
Sync[F].raiseError(
new FileNotFoundException(s"No docspell-addon.{yaml|json} file found in $dir!")
)
)
}
def findInZip[F[_]: Async](zipFile: Stream[F, Byte]): F[AddonMeta] = {
val fail: F[AddonMeta] = Async[F].raiseError(
new FileNotFoundException(
s"No docspell-addon.{yaml|json} file found in zip!"
)
)
zipFile
.through(Zip.unzip(8192, Glob("**/docspell-addon.*")))
.filter(bin => !bin.name.endsWith("/"))
.flatMap { bin =>
if (bin.extensionIn(Set("json"))) Stream.eval(AddonMeta.fromJsonBytes(bin.data))
else if (bin.extensionIn(Set("yaml", "yml")))
Stream.eval(AddonMeta.fromYamlBytes(bin.data))
else Stream.empty
}
.take(1)
.compile
.last
.flatMap(_.map(Sync[F].pure).getOrElse(fail))
}
}

View File

@ -0,0 +1,9 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.addons
case class AddonRef(archive: AddonArchive, args: String)

View File

@ -0,0 +1,114 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.addons
import cats.Monoid
import docspell.addons.out.AddonOutput
import io.circe.generic.extras.Configuration
import io.circe.generic.extras.semiauto.{deriveConfiguredDecoder, deriveConfiguredEncoder}
import io.circe.generic.semiauto.{deriveDecoder, deriveEncoder}
import io.circe.{Codec, Decoder, Encoder}
sealed trait AddonResult {
def toEither: Either[Throwable, AddonOutput]
def isSuccess: Boolean = toEither.isRight
def isFailure: Boolean = !isSuccess
def cast: AddonResult = this
}
object AddonResult {
/** The addon was run successful, but decoding its stdout failed. */
case class DecodingError(message: String) extends AddonResult {
def toEither = Left(new IllegalStateException(message))
}
object DecodingError {
implicit val jsonEncoder: Encoder[DecodingError] = deriveEncoder
implicit val jsonDecoder: Decoder[DecodingError] = deriveDecoder
}
def decodingError(message: String): AddonResult =
DecodingError(message)
def decodingError(ex: Throwable): AddonResult =
DecodingError(ex.getMessage)
/** Running the addon resulted in an invalid return code (!= 0). */
case class ExecutionError(rc: Int) extends AddonResult {
def toEither = Left(new IllegalStateException(s"Exit code: $rc"))
}
object ExecutionError {
implicit val jsonEncoder: Encoder[ExecutionError] = deriveEncoder
implicit val jsonDecoder: Decoder[ExecutionError] = deriveDecoder
}
def executionError(rc: Int): AddonResult =
ExecutionError(rc)
/** The execution of the addon failed with an exception. */
case class ExecutionFailed(error: Throwable) extends AddonResult {
def toEither = Left(error)
}
object ExecutionFailed {
implicit val throwableCodec: Codec[Throwable] =
Codec.from(
Decoder[String].emap(str => Right(ErrorMessageThrowable(str))),
Encoder[String].contramap(_.getMessage)
)
implicit val jsonEncoder: Encoder[ExecutionFailed] = deriveEncoder
implicit val jsonDecoder: Decoder[ExecutionFailed] = deriveDecoder
private class ErrorMessageThrowable(msg: String) extends RuntimeException(msg) {
override def fillInStackTrace() = this
}
private object ErrorMessageThrowable {
def apply(str: String): Throwable = new ErrorMessageThrowable(str)
}
}
def executionFailed(error: Throwable): AddonResult =
ExecutionFailed(error)
/** The addon was run successfully and its output was decoded (if any). */
case class Success(output: AddonOutput) extends AddonResult {
def toEither = Right(output)
}
object Success {
implicit val jsonEncoder: Encoder[Success] = deriveEncoder
implicit val jsonDecoder: Decoder[Success] = deriveDecoder
}
def success(output: AddonOutput): AddonResult =
Success(output)
val empty: AddonResult = Success(AddonOutput.empty)
def combine(a: AddonResult, b: AddonResult): AddonResult =
(a, b) match {
case (Success(o1), Success(o2)) => Success(AddonOutput.combine(o1, o2))
case (Success(_), e) => e
case (e, Success(_)) => e
case _ => a
}
implicit val deriveConfig: Configuration =
Configuration.default.withDiscriminator("result").withKebabCaseConstructorNames
implicit val jsonDecoder: Decoder[AddonResult] = deriveConfiguredDecoder
implicit val jsonEncoder: Encoder[AddonResult] = deriveConfiguredEncoder
implicit val addonResultMonoid: Monoid[AddonResult] =
Monoid.instance(empty, combine)
}

View File

@ -0,0 +1,110 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.addons
import cats.Applicative
import cats.effect._
import cats.syntax.all._
import fs2.Stream
import docspell.addons.runner._
import docspell.common.exec.Env
import docspell.logging.Logger
trait AddonRunner[F[_]] {
def runnerType: List[RunnerType]
def run(
logger: Logger[F],
env: Env,
ctx: Context
): F[AddonResult]
}
object AddonRunner {
def forType[F[_]: Async](cfg: AddonExecutorConfig)(rt: RunnerType) =
rt match {
case RunnerType.NixFlake => NixFlakeRunner[F](cfg)
case RunnerType.Docker => DockerRunner[F](cfg)
case RunnerType.Trivial => TrivialRunner[F](cfg)
}
def failWith[F[_]](errorMsg: String)(implicit F: Applicative[F]): AddonRunner[F] =
pure(AddonResult.executionFailed(new Exception(errorMsg)))
def pure[F[_]: Applicative](result: AddonResult): AddonRunner[F] =
new AddonRunner[F] {
val runnerType = Nil
def run(logger: Logger[F], env: Env, ctx: Context) =
Applicative[F].pure(result)
}
def firstSuccessful[F[_]: Sync](runners: List[AddonRunner[F]]): AddonRunner[F] =
runners match {
case Nil => failWith("No runner available!")
case a :: Nil => a
case _ =>
new AddonRunner[F] {
val runnerType = runners.flatMap(_.runnerType).distinct
def run(logger: Logger[F], env: Env, ctx: Context) =
Stream
.emits(runners)
.evalTap(r =>
logger.info(
s"Attempt to run addon ${ctx.meta.nameAndVersion} with runner ${r.runnerType}"
)
)
.evalMap(_.run(logger, env, ctx))
.flatMap {
case r @ AddonResult.Success(_) => Stream.emit(r.cast.some)
case r @ AddonResult.ExecutionFailed(ex) =>
if (ctx.meta.isPure) {
logger.stream
.warn(ex)(s"Addon runner failed, try next.")
.as(r.cast.some)
} else {
logger.stream.warn(ex)(s"Addon runner failed!").as(None)
}
case r @ AddonResult.ExecutionError(rc) =>
if (ctx.meta.isPure) {
logger.stream
.warn(s"Addon runner returned non-zero: $rc. Try next.")
.as(r.cast.some)
} else {
logger.stream.warn(s"Addon runner returned non-zero: $rc!").as(None)
}
case AddonResult.DecodingError(message) =>
// Don't retry as it is very unlikely that the output differs using another runner
// This is most likely a bug in the addon
logger.stream
.warn(
s"Error decoding the output of the addon ${ctx.meta.nameAndVersion}: $message. Stopping here. This is likely a bug in the addon."
)
.as(None)
}
.unNoneTerminate
.takeThrough(_.isFailure)
.compile
.last
.flatMap {
case Some(r) => r.pure[F]
case None =>
AddonResult
.executionFailed(new NoSuchElementException("No runner left :("))
.pure[F]
}
}
}
def firstSuccessful[F[_]: Sync](
runner: AddonRunner[F],
runners: AddonRunner[F]*
): AddonRunner[F] =
firstSuccessful(runner :: runners.toList)
}

View File

@ -0,0 +1,55 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.addons
import cats.data.NonEmptyList
import io.circe.{Decoder, Encoder}
sealed trait AddonTriggerType {
def name: String
}
object AddonTriggerType {
/** The final step when processing an item. */
case object FinalProcessItem extends AddonTriggerType {
val name = "final-process-item"
}
/** The final step when reprocessing an item. */
case object FinalReprocessItem extends AddonTriggerType {
val name = "final-reprocess-item"
}
/** Running periodically based on a schedule. */
case object Scheduled extends AddonTriggerType {
val name = "scheduled"
}
/** Running (manually) on some existing item. */
case object ExistingItem extends AddonTriggerType {
val name = "existing-item"
}
val all: NonEmptyList[AddonTriggerType] =
NonEmptyList.of(FinalProcessItem, FinalReprocessItem, Scheduled, ExistingItem)
def fromString(str: String): Either[String, AddonTriggerType] =
all
.find(e => e.name.equalsIgnoreCase(str))
.toRight(s"Invalid addon trigger type: $str")
def unsafeFromString(str: String): AddonTriggerType =
fromString(str).fold(sys.error, identity)
implicit val jsonEncoder: Encoder[AddonTriggerType] =
Encoder.encodeString.contramap(_.name)
implicit val jsonDecoder: Decoder[AddonTriggerType] =
Decoder.decodeString.emap(fromString)
}

View File

@ -0,0 +1,72 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.addons
import fs2.io.file.Path
import docspell.common._
import docspell.common.exec.{Args, Env, SysCmd}
/** Context a list of addons is executed in.
*
* Each addon has its own `addonDir`, but all share the same `baseDir` in one run.
*/
case class Context(
addon: AddonRef,
meta: AddonMeta,
baseDir: Path,
addonDir: Path,
outputDir: Path,
cacheDir: Path
) {
def userInputFile = Context.userInputFile(baseDir)
def tempDir = Context.tempDir(baseDir)
private[addons] def addonCommand(
binary: String,
timeout: Duration,
relativeToBase: Boolean,
outputDir: Option[String],
cacheDir: Option[String]
): SysCmd = {
val execBin = Option
.when(relativeToBase)(binary)
.getOrElse((baseDir / binary).toString)
val input = Option
.when(relativeToBase)(baseDir.relativize(userInputFile))
.getOrElse(userInputFile)
val allArgs =
Args(meta.args.getOrElse(Nil)).append(input)
val envAddonDir = Option
.when(relativeToBase)(baseDir.relativize(addonDir))
.getOrElse(addonDir)
val envTmpDir = Option
.when(relativeToBase)(baseDir.relativize(tempDir))
.getOrElse(tempDir)
val outDir = outputDir.getOrElse(this.outputDir.toString)
val cache = cacheDir.getOrElse(this.cacheDir.toString)
val moreEnv =
Env.of(
"ADDON_DIR" -> envAddonDir.toString,
"TMPDIR" -> envTmpDir.toString,
"TMP_DIR" -> envTmpDir.toString,
"OUTPUT_DIR" -> outDir,
"CACHE_DIR" -> cache
)
SysCmd(execBin, allArgs).withTimeout(timeout).addEnv(moreEnv)
}
}
object Context {
def userInputFile(base: Path): Path =
base / "arguments" / "user-input"
def tempDir(base: Path): Path =
base / "temp"
}

View File

@ -0,0 +1,74 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.addons
import cats.effect._
import cats.syntax.all._
import cats.{Applicative, Monad}
import fs2.io.file.{Files, Path, PosixPermissions}
object Directory {
def create[F[_]: Files: Applicative](dir: Path): F[Path] =
Files[F]
.createDirectories(dir, PosixPermissions.fromOctal("777"))
.as(dir)
def createAll[F[_]: Files: Applicative](dir: Path, dirs: Path*): F[Unit] =
(dir :: dirs.toList).traverse_(Files[F].createDirectories(_))
def nonEmpty[F[_]: Files: Sync](dir: Path): F[Boolean] =
List(
Files[F].isDirectory(dir),
Files[F].list(dir).take(1).compile.last.map(_.isDefined)
).sequence.map(_.forall(identity))
def isEmpty[F[_]: Files: Sync](dir: Path): F[Boolean] =
nonEmpty(dir).map(b => !b)
def temp[F[_]: Files](parent: Path, prefix: String): Resource[F, Path] =
for {
_ <- Resource.eval(Files[F].createDirectories(parent))
d <- mkTemp(parent, prefix)
} yield d
def temp2[F[_]: Files](
parent: Path,
prefix1: String,
prefix2: String
): Resource[F, (Path, Path)] =
for {
_ <- Resource.eval(Files[F].createDirectories(parent))
a <- mkTemp(parent, prefix1)
b <- mkTemp(parent, prefix2)
} yield (a, b)
def createTemp[F[_]: Files: Monad](
parent: Path,
prefix: String
): F[Path] =
for {
_ <- Files[F].createDirectories(parent)
d <- mkTemp_(parent, prefix)
} yield d
private def mkTemp[F[_]: Files](parent: Path, prefix: String): Resource[F, Path] =
Files[F]
.tempDirectory(
parent.some,
prefix,
PosixPermissions.fromOctal("777")
)
private def mkTemp_[F[_]: Files](parent: Path, prefix: String): F[Path] =
Files[F]
.createTempDirectory(
parent.some,
prefix,
PosixPermissions.fromOctal("777")
)
}

View File

@ -0,0 +1,32 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.addons
import cats.effect.Resource
import fs2.io.file.{Files, Path}
import docspell.common.exec.Env
case class InputEnv(
addons: List[AddonRef],
baseDir: Path,
outputDir: Path,
cacheDir: Path,
env: Env
) {
def addEnv(key: String, value: String): InputEnv =
copy(env = env.add(key, value))
def addEnv(vp: (String, String)*): InputEnv =
copy(env = env.addAll(vp.toMap))
def addEnv(vm: Map[String, String]): InputEnv =
copy(env = env ++ Env(vm))
def withTempBase[F[_]: Files]: Resource[F, InputEnv] =
Directory.temp(baseDir, "addon-").map(path => copy(baseDir = path))
}

View File

@ -0,0 +1,43 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.addons
import cats.Monad
import cats.data.Kleisli
import cats.effect.kernel.Sync
import cats.syntax.all._
import fs2.io.file.Files
trait Middleware[F[_]] extends (AddonExec[F] => AddonExec[F]) { self =>
def >>(next: Middleware[F]): Middleware[F] =
Middleware(self.andThen(next))
}
object Middleware {
def apply[F[_]](f: AddonExec[F] => AddonExec[F]): Middleware[F] =
a => f(a)
def identity[F[_]]: Middleware[F] = Middleware(scala.Predef.identity)
/** Uses a temporary base dir that is removed after execution. Use this as the last
* layer!
*/
def ephemeralRun[F[_]: Files: Sync]: Middleware[F] =
Middleware(a => Kleisli(_.withTempBase.use(a.run)))
/** Prepare running an addon */
def prepare[F[_]: Monad](
prep: Kleisli[F, InputEnv, InputEnv]
): Middleware[F] =
Middleware(a => Kleisli(in => prep.run(in).flatMap(a.run)))
def postProcess[F[_]: Monad](
post: Kleisli[F, AddonExecutionResult, Unit]
): Middleware[F] =
Middleware(_.flatMapF(r => post.map(_ => r).run(r)))
}

View File

@ -0,0 +1,69 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.addons
import cats.data.NonEmptyList
import cats.syntax.all._
import io.circe.{Decoder, Encoder}
sealed trait RunnerType {
def name: String
def fold[A](
nixFlake: => A,
docker: => A,
trivial: => A
): A
}
object RunnerType {
case object NixFlake extends RunnerType {
val name = "nix-flake"
def fold[A](
nixFlake: => A,
docker: => A,
trivial: => A
): A = nixFlake
}
case object Docker extends RunnerType {
val name = "docker"
def fold[A](
nixFlake: => A,
docker: => A,
trivial: => A
): A = docker
}
case object Trivial extends RunnerType {
val name = "trivial"
def fold[A](
nixFlake: => A,
docker: => A,
trivial: => A
): A = trivial
}
val all: NonEmptyList[RunnerType] =
NonEmptyList.of(NixFlake, Docker, Trivial)
def fromString(str: String): Either[String, RunnerType] =
all.find(_.name.equalsIgnoreCase(str)).toRight(s"Invalid runner value: $str")
def unsafeFromString(str: String): RunnerType =
fromString(str).fold(sys.error, identity)
def fromSeparatedString(str: String): Either[String, List[RunnerType]] =
str.split("[\\s,]+").toList.map(_.trim).traverse(fromString)
implicit val jsonDecoder: Decoder[RunnerType] =
Decoder[String].emap(RunnerType.fromString)
implicit val jsonEncoder: Encoder[RunnerType] =
Encoder[String].contramap(_.name)
}

View File

@ -0,0 +1,44 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.addons.out
import cats.kernel.Monoid
import docspell.common.bc.BackendCommand
import io.circe.generic.extras.Configuration
import io.circe.generic.extras.semiauto.{deriveConfiguredDecoder, deriveConfiguredEncoder}
import io.circe.{Decoder, Encoder}
/** Decoded stdout result from executing an addon. */
case class AddonOutput(
commands: List[BackendCommand] = Nil,
files: List[ItemFile] = Nil,
newItems: List[NewItem] = Nil
)
object AddonOutput {
val empty: AddonOutput = AddonOutput()
def combine(a: AddonOutput, b: AddonOutput): AddonOutput =
AddonOutput(a.commands ++ b.commands, a.files ++ b.files)
implicit val addonResultMonoid: Monoid[AddonOutput] =
Monoid.instance(empty, combine)
implicit val jsonConfig: Configuration =
Configuration.default.withDefaults
implicit val jsonDecoder: Decoder[AddonOutput] = deriveConfiguredDecoder
implicit val jsonEncoder: Encoder[AddonOutput] = deriveConfiguredEncoder
def fromString(str: String): Either[Throwable, AddonOutput] =
io.circe.parser.decode[AddonOutput](str)
def unsafeFromString(str: String): AddonOutput =
fromString(str).fold(throw _, identity)
}

View File

@ -0,0 +1,102 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.addons.out
import cats.data.OptionT
import cats.effect._
import cats.syntax.all._
import fs2.io.file.{Files, Path}
import docspell.common._
import docspell.files.FileSupport._
import docspell.logging.Logger
import io.circe.generic.extras.Configuration
import io.circe.generic.extras.semiauto.{deriveConfiguredDecoder, deriveConfiguredEncoder}
import io.circe.{Decoder, Encoder}
/** Addons can produce files in their output directory. These can be named here in order
* to do something with them.
*
* - textFiles will replace the extracted text with the contents of the file
* - pdfFiles will add/replace the converted pdf with the given file
* - previewImages will add/replace preview images
* - newFiles will be added as new attachments to the item
*
* Files must be referenced by attachment id.
*/
final case class ItemFile(
itemId: Ident,
textFiles: Map[String, String] = Map.empty,
pdfFiles: Map[String, String] = Map.empty,
previewImages: Map[String, String] = Map.empty,
newFiles: List[NewFile] = Nil
) {
def isEmpty: Boolean =
textFiles.isEmpty && pdfFiles.isEmpty && previewImages.isEmpty
def nonEmpty: Boolean = !isEmpty
def resolveTextFiles[F[_]: Files: Sync](
logger: Logger[F],
outputDir: Path
): F[List[(String, Path)]] =
resolveFiles(logger, outputDir, MimeType.text("*"), textFiles)
def resolvePdfFiles[F[_]: Files: Sync](
logger: Logger[F],
outputDir: Path
): F[List[(String, Path)]] =
resolveFiles(logger, outputDir, MimeType.pdf, pdfFiles)
def resolvePreviewFiles[F[_]: Files: Sync](
logger: Logger[F],
outputDir: Path
): F[List[(String, Path)]] =
resolveFiles(logger, outputDir, MimeType.image("*"), previewImages)
def resolveNewFiles[F[_]: Files: Sync](
logger: Logger[F],
outputDir: Path
): F[List[(NewFile, Path)]] =
newFiles.traverseFilter(nf =>
nf.resolveFile(logger, outputDir).map(_.map(p => (nf, p)))
)
private def resolveFiles[F[_]: Files: Sync](
logger: Logger[F],
outputDir: Path,
mime: MimeType,
files: Map[String, String]
): F[List[(String, Path)]] = {
val allFiles =
files.toList.map(t => t._1 -> outputDir / t._2)
allFiles.traverseFilter { case (key, file) =>
OptionT(file.detectMime)
.flatMapF(fileType =>
if (mime.matches(fileType)) (key -> file).some.pure[F]
else
logger
.warn(
s"File $file provided as ${mime.asString} file, but was recognized as ${fileType.asString}. Ignoring it."
)
.as(None: Option[(String, Path)])
)
.value
}
}
}
object ItemFile {
implicit val jsonConfig: Configuration =
Configuration.default.withDefaults
implicit val jsonEncoder: Encoder[ItemFile] = deriveConfiguredEncoder
implicit val jsonDecoder: Decoder[ItemFile] = deriveConfiguredDecoder
}

View File

@ -0,0 +1,77 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.addons.out
import cats.effect.Sync
import cats.syntax.all._
import fs2.io.file.{Files, Path}
import docspell.addons.out.NewFile.Meta
import docspell.common.ProcessItemArgs.ProcessMeta
import docspell.common.{Ident, Language}
import docspell.logging.Logger
import io.circe.Codec
import io.circe.generic.extras.Configuration
import io.circe.generic.extras.semiauto.deriveConfiguredCodec
import io.circe.generic.semiauto.deriveCodec
case class NewFile(metadata: Meta = Meta.empty, file: String) {
def resolveFile[F[_]: Files: Sync](
logger: Logger[F],
outputDir: Path
): F[Option[Path]] = {
val target = outputDir / file
Files[F]
.exists(target)
.flatMap(flag =>
if (flag) target.some.pure[F]
else logger.warn(s"File not found: $file").as(Option.empty)
)
}
}
object NewFile {
case class Meta(
language: Option[Language],
skipDuplicate: Option[Boolean],
attachmentsOnly: Option[Boolean]
) {
def toProcessMeta(
cid: Ident,
itemId: Ident,
collLang: Option[Language],
sourceAbbrev: String
): ProcessMeta =
ProcessMeta(
collective = cid,
itemId = Some(itemId),
language = language.orElse(collLang).getOrElse(Language.English),
direction = None,
sourceAbbrev = sourceAbbrev,
folderId = None,
validFileTypes = Seq.empty,
skipDuplicate = skipDuplicate.getOrElse(true),
fileFilter = None,
tags = None,
reprocess = false,
attachmentsOnly = attachmentsOnly
)
}
object Meta {
val empty = Meta(None, None, None)
implicit val jsonCodec: Codec[Meta] = deriveCodec
}
implicit val jsonConfig: Configuration = Configuration.default.withDefaults
implicit val jsonCodec: Codec[NewFile] = deriveConfiguredCodec
}

View File

@ -0,0 +1,92 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.addons.out
import cats.Monad
import cats.syntax.all._
import fs2.io.file.{Files, Path}
import docspell.addons.out.NewItem.Meta
import docspell.common._
import docspell.logging.Logger
import io.circe.generic.semiauto.{deriveDecoder, deriveEncoder}
import io.circe.{Decoder, Encoder}
case class NewItem(metadata: Option[Meta], files: List[String]) {
def toProcessMeta(
cid: Ident,
collLang: Option[Language],
sourceAbbrev: String
): ProcessItemArgs.ProcessMeta =
metadata
.getOrElse(Meta(None, None, None, None, None, None, None))
.toProcessArgs(cid, collLang, sourceAbbrev)
def resolveFiles[F[_]: Files: Monad](
logger: Logger[F],
outputDir: Path
): F[List[Path]] = {
val allFiles =
files.map(name => outputDir / name)
allFiles.traverseFilter { file =>
Files[F]
.exists(file)
.flatMap {
case true => file.some.pure[F]
case false =>
logger
.warn(s"File $file doesn't exist. Ignoring it.")
.as(None)
}
}
}
}
object NewItem {
case class Meta(
language: Option[Language],
direction: Option[Direction],
folderId: Option[Ident],
source: Option[String],
skipDuplicate: Option[Boolean],
tags: Option[List[String]],
attachmentsOnly: Option[Boolean]
) {
def toProcessArgs(
cid: Ident,
collLang: Option[Language],
sourceAbbrev: String
): ProcessItemArgs.ProcessMeta =
ProcessItemArgs.ProcessMeta(
collective = cid,
itemId = None,
language = language.orElse(collLang).getOrElse(Language.English),
direction = direction,
sourceAbbrev = source.getOrElse(sourceAbbrev),
folderId = folderId,
validFileTypes = Seq.empty,
skipDuplicate = skipDuplicate.getOrElse(true),
fileFilter = None,
tags = tags,
reprocess = false,
attachmentsOnly = attachmentsOnly
)
}
object Meta {
implicit val jsonEncoder: Encoder[Meta] = deriveEncoder
implicit val jsonDecoder: Decoder[Meta] = deriveDecoder
}
implicit val jsonDecoder: Decoder[NewItem] = deriveDecoder
implicit val jsonEncoder: Encoder[NewItem] = deriveEncoder
}

View File

@ -0,0 +1,15 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell
import cats.data.Kleisli
package object addons {
type AddonExec[F[_]] = Kleisli[F, InputEnv, AddonExecutionResult]
}

View File

@ -0,0 +1,43 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.addons.runner
import cats.Applicative
import cats.effect.{Ref, Sync}
import cats.syntax.all._
import fs2.Pipe
trait CollectOut[F[_]] {
def get: F[String]
def append: Pipe[F, String, String]
}
object CollectOut {
def none[F[_]: Applicative]: CollectOut[F] =
new CollectOut[F] {
def get = "".pure[F]
def append = identity
}
def buffer[F[_]: Sync]: F[CollectOut[F]] =
Ref
.of[F, Vector[String]](Vector.empty)
.map(buffer =>
new CollectOut[F] {
override def get =
buffer.get.map(_.mkString("\n").trim)
override def append =
_.evalTap(line =>
if (line.trim.nonEmpty) buffer.update(_.appended(line)) else ().pure[F]
)
}
)
}

View File

@ -0,0 +1,82 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.addons.runner
import fs2.io.file.Path
import docspell.common.Duration
import docspell.common.exec.{Args, Env, SysCmd}
/** Builder for a docker system command. */
case class DockerBuilder(
dockerBinary: String,
subCmd: String,
timeout: Duration,
containerName: Option[String] = None,
env: Env = Env.empty,
mounts: Args = Args.empty,
network: Option[String] = Some("host"),
workingDir: Option[String] = None,
imageName: Option[String] = None,
cntCmd: Args = Args.empty
) {
def containerCmd(args: Args): DockerBuilder =
copy(cntCmd = args)
def containerCmd(args: Seq[String]): DockerBuilder =
copy(cntCmd = Args(args))
def imageName(name: String): DockerBuilder =
copy(imageName = Some(name))
def workDirectory(dir: String): DockerBuilder =
copy(workingDir = Some(dir))
def withDockerBinary(bin: String): DockerBuilder =
copy(dockerBinary = bin)
def withSubCmd(cmd: String): DockerBuilder =
copy(subCmd = cmd)
def withEnv(key: String, value: String): DockerBuilder =
copy(env = env.add(key, value))
def withEnv(moreEnv: Env): DockerBuilder =
copy(env = env ++ moreEnv)
def privateNetwork(flag: Boolean): DockerBuilder =
if (flag) copy(network = Some("none"))
else copy(network = Some("host"))
def mount(
hostDir: Path,
cntDir: Option[String] = None,
readOnly: Boolean = true
): DockerBuilder = {
val target = cntDir.getOrElse(hostDir.toString)
val ro = Option.when(readOnly)(",readonly").getOrElse("")
val opt = s"type=bind,source=$hostDir,target=$target${ro}"
copy(mounts = mounts.append("--mount", opt))
}
def withName(containerName: String): DockerBuilder =
copy(containerName = Some(containerName))
def build: SysCmd =
SysCmd(dockerBinary, buildArgs).withTimeout(timeout)
private def buildArgs: Args =
Args
.of(subCmd)
.append("--rm")
.option("--name", containerName)
.append(mounts)
.option("--network", network)
.append(env.mapConcat((k, v) => List("--env", s"${k}=${v}")))
.option("-w", workingDir)
.appendOpt(imageName)
.append(cntCmd)
}

View File

@ -0,0 +1,93 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.addons.runner
import cats.data.OptionT
import cats.effect._
import cats.syntax.all._
import docspell.addons.AddonExecutorConfig.DockerConfig
import docspell.addons._
import docspell.common.Duration
import docspell.common.exec.{Env, SysCmd, SysExec}
import docspell.common.util.Random
import docspell.logging.Logger
final class DockerRunner[F[_]: Async](cfg: DockerRunner.Config) extends AddonRunner[F] {
val runnerType = List(RunnerType.Docker)
def run(
logger: Logger[F],
env: Env,
ctx: Context
) = for {
_ <- OptionT.whenF(requireBuild(ctx))(build(logger, ctx)).value
suffix <- Random[F].string(4)
cmd = createDockerCommand(env, ctx, suffix)
result <- RunnerUtil.runAddonCommand(logger, cmd, ctx)
} yield result
def createDockerCommand(
env: Env,
ctx: Context,
suffix: String
): SysCmd = {
val outputPath = "/mnt/output"
val cachePath = "/mnt/cache"
val addonArgs =
ctx.addonCommand(
"",
Duration.zero,
relativeToBase = true,
outputPath.some,
cachePath.some
)
DockerBuilder(cfg.docker.dockerBinary, "run", cfg.timeout)
.withName(ctx.meta.nameAndVersion + "-" + suffix)
.withEnv(env)
.withEnv(addonArgs.env)
.mount(ctx.baseDir, "/mnt/work".some, readOnly = false)
.mount(ctx.outputDir, outputPath.some, readOnly = false)
.mount(ctx.cacheDir, cachePath.some, readOnly = false)
.workDirectory("/mnt/work")
.privateNetwork(ctx.meta.isPure)
.imageName(imageName(ctx))
.containerCmd(addonArgs.args)
.build
}
def build(logger: Logger[F], ctx: Context): F[Unit] =
for {
_ <- logger.info(s"Building docker image for addon ${ctx.meta.nameAndVersion}")
cmd = cfg.docker.dockerBuild(imageName(ctx))
_ <- SysExec(cmd, logger, ctx.addonDir.some)
.flatMap(_.logOutputs(logger, "docker build"))
.use(_.waitFor())
_ <- logger.info(s"Docker image built successfully")
} yield ()
private def requireBuild(ctx: Context) =
ctx.meta.runner
.flatMap(_.docker)
.flatMap(_.image)
.isEmpty
private def imageName(ctx: Context): String =
ctx.meta.runner
.flatMap(_.docker)
.flatMap(_.image)
.getOrElse(s"${ctx.meta.meta.name}:latest")
}
object DockerRunner {
def apply[F[_]: Async](cfg: AddonExecutorConfig): DockerRunner[F] =
new DockerRunner[F](Config(cfg.dockerRunner, cfg.runTimeout))
case class Config(docker: DockerConfig, timeout: Duration)
}

View File

@ -0,0 +1,83 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.addons.runner
import fs2.io.file.Path
import docspell.common.exec.{Args, Env, SysCmd}
case class NSpawnBuilder(
child: SysCmd,
chroot: Path,
spawnBinary: String = "systemd-nspawn",
sudoBinary: String = "sudo",
args: Args = Args.empty,
env: Env = Env.empty
) {
def withNSpawnBinary(bin: String): NSpawnBuilder =
copy(spawnBinary = bin)
def withSudoBinary(bin: String): NSpawnBuilder =
copy(sudoBinary = bin)
def withEnv(key: String, value: String): NSpawnBuilder =
copy(args = args.append(s"--setenv=$key=$value"))
def withEnvOpt(key: String, value: Option[String]): NSpawnBuilder =
value.map(v => withEnv(key, v)).getOrElse(this)
def withName(containerName: String): NSpawnBuilder =
copy(args = args.append(s"--machine=$containerName"))
def mount(
hostDir: Path,
cntDir: Option[String] = None,
readOnly: Boolean = true
): NSpawnBuilder = {
val bind = if (readOnly) "--bind-ro" else "--bind"
val target = cntDir.map(dir => s":$dir").getOrElse("")
copy(args = args.append(s"${bind}=${hostDir}${target}"))
}
def workDirectory(dir: String): NSpawnBuilder =
copy(args = args.append(s"--chdir=$dir"))
def portMap(port: Int): NSpawnBuilder =
copy(args = args.append("-p", port.toString))
def privateNetwork(flag: Boolean): NSpawnBuilder =
if (flag) copy(args = args.append("--private-network"))
else this
def build: SysCmd =
SysCmd(
program = if (sudoBinary.nonEmpty) sudoBinary else spawnBinary,
args = buildArgs,
timeout = child.timeout,
env = env
)
private def buildArgs: Args =
Args
.of("--private-users=identity") // can't use -U because need writeable bind mounts
.append("--notify-ready=yes")
.append("--ephemeral")
.append("--as-pid2")
.append("--console=pipe")
.append("--no-pager")
.append("--bind-ro=/bin")
.append("--bind-ro=/usr/bin")
.append("--bind-ro=/nix/store")
.append(s"--directory=$chroot")
.append(args)
.append(child.env.map((n, v) => s"--setenv=$n=$v"))
.prependWhen(sudoBinary.nonEmpty)(spawnBinary)
.append("--")
.append(child.program)
.append(child.args)
}

View File

@ -0,0 +1,123 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.addons.runner
import cats.effect._
import cats.syntax.all._
import fs2.Stream
import fs2.io.file.{Files, Path}
import docspell.addons.AddonExecutorConfig.{NSpawn, NixConfig}
import docspell.addons._
import docspell.addons.runner.NixFlakeRunner.PreCtx
import docspell.common.Duration
import docspell.common.exec._
import docspell.logging.Logger
final class NixFlakeRunner[F[_]: Async](cfg: NixFlakeRunner.Config)
extends AddonRunner[F] {
val runnerType = List(RunnerType.NixFlake)
def run(
logger: Logger[F],
env: Env,
ctx: Context
): F[AddonResult] =
prepare(logger, ctx)
.flatMap { preCtx =>
if (preCtx.nspawnEnabled) runInContainer(logger, env, preCtx, ctx)
else runOnHost(logger, env, preCtx, ctx)
}
def prepare(logger: Logger[F], ctx: Context): F[PreCtx] =
for {
_ <- logger.info(s"Prepare addon ${ctx.meta.nameAndVersion} for executing via nix")
_ <- logger.debug(s"Building with nix build")
_ <- SysExec(cfg.nixBuild, logger, workdir = ctx.addonDir.some)
.flatMap(_.logOutputs(logger, "nix build"))
.use(_.waitFor())
bin <- findFile(ctx.addonDir / "result" / "bin", ctx.addonDir / "result")
_ <- logger.debug(s"Build done, found binary: $bin")
_ <- logger.debug(s"Checking for systemd-nspawn…")
cnt <- checkContainer(logger)
_ <-
if (cnt)
logger.debug(s"Using systemd-nspawn to run addon in a container.")
else
logger.info(s"Running via systemd-nspawn is disabled in the config file")
} yield PreCtx(cnt, ctx.baseDir.relativize(bin))
private def checkContainer(logger: Logger[F]): F[Boolean] =
if (!cfg.nspawn.enabled) false.pure[F]
else RunnerUtil.checkContainer(logger, cfg.nspawn)
private def runOnHost(
logger: Logger[F],
env: Env,
preCtx: PreCtx,
ctx: Context
): F[AddonResult] = {
val cmd =
SysCmd(preCtx.binary.toString, Args.empty).withTimeout(cfg.timeout).addEnv(env)
RunnerUtil.runDirectly(logger, ctx)(cmd)
}
private def runInContainer(
logger: Logger[F],
env: Env,
preCtx: PreCtx,
ctx: Context
): F[AddonResult] = {
val cmd = SysCmd(preCtx.binary.toString, Args.empty)
.withTimeout(cfg.timeout)
.addEnv(env)
RunnerUtil.runInContainer(logger, cfg.nspawn, ctx)(cmd)
}
/** Find first file, try directories in given order. */
private def findFile(firstDir: Path, more: Path*): F[Path] = {
val fail: F[Path] = Sync[F].raiseError(
new NoSuchElementException(
s"No file found to execute in ${firstDir :: more.toList}"
)
)
Stream
.emits(more)
.cons1(firstDir)
.flatMap(dir =>
Files[F]
.list(dir)
.evalFilter(p => Files[F].isDirectory(p).map(!_))
.take(1)
)
.take(1)
.compile
.last
.flatMap(_.fold(fail)(Sync[F].pure))
}
}
object NixFlakeRunner {
def apply[F[_]: Async](cfg: AddonExecutorConfig): NixFlakeRunner[F] =
new NixFlakeRunner[F](Config(cfg.nixRunner, cfg.nspawn, cfg.runTimeout))
case class Config(
nix: NixConfig,
nspawn: NSpawn,
timeout: Duration
) {
val nixBuild =
SysCmd(nix.nixBinary, Args.of("build")).withTimeout(nix.buildTimeout)
val nspawnVersion = nspawn.nspawnVersion
}
case class PreCtx(nspawnEnabled: Boolean, binary: Path)
}

View File

@ -0,0 +1,171 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.addons.runner
import cats.data.OptionT
import cats.effect.{Async, Sync}
import cats.syntax.all._
import fs2.Pipe
import fs2.io.file.Files
import docspell.addons._
import docspell.addons.out.AddonOutput
import docspell.common.exec.{SysCmd, SysExec}
import docspell.common.util.Random
import docspell.logging.Logger
import io.circe.{parser => JsonParser}
private[addons] object RunnerUtil {
/** Run the given `cmd` on this machine.
*
* The `cmd` is containing a template command to execute the addon. The path are
* expected to be relative to the `ctx.baseDir`. Additional arguments and environment
* variables are added as configured in the addon.
*/
def runDirectly[F[_]: Async](
logger: Logger[F],
ctx: Context
)(cmd: SysCmd): F[AddonResult] = {
val addonCmd = ctx
.addonCommand(cmd.program, cmd.timeout, relativeToBase = false, None, None)
.withArgs(_.append(cmd.args))
.addEnv(cmd.env)
runAddonCommand(logger, addonCmd, ctx)
}
/** Run the given `cmd` inside a container via systemd-nspawn.
*
* The `cmd` is containing a template command to execute the addon. The path are
* expected to be relative to the `ctx.baseDir`. Additional arguments and environment
* variables are added as configured in the addon.
*/
def runInContainer[F[_]: Async](
logger: Logger[F],
cfg: AddonExecutorConfig.NSpawn,
ctx: Context
)(cmd: SysCmd): F[AddonResult] = {
val outputPath = "/mnt/output"
val cachePath = "/mnt/cache"
val addonCmd = ctx
.addonCommand(
cmd.program,
cmd.timeout,
relativeToBase = true,
outputPath.some,
cachePath.some
)
.withArgs(_.append(cmd.args))
.addEnv(cmd.env)
val chroot = ctx.baseDir / "cnt-root"
val nspawn = NSpawnBuilder(addonCmd, chroot)
.withNSpawnBinary(cfg.nspawnBinary)
.withSudoBinary(cfg.sudoBinary)
.mount(ctx.baseDir, "/mnt/work".some, readOnly = false)
.mount(ctx.cacheDir, cachePath.some, readOnly = false)
.mount(ctx.outputDir, outputPath.some, readOnly = false)
.workDirectory("/mnt/work")
.withEnv("XDG_RUNTIME_DIR", "/mnt/work")
.privateNetwork(ctx.meta.isPure)
for {
suffix <- Random[F].string(4)
_ <- List(chroot).traverse_(Files[F].createDirectories)
res <- runAddonCommand(
logger,
nspawn.withName(ctx.meta.nameAndVersion + "-" + suffix).build,
ctx
)
// allow some time to unregister the current container
// only important when same addons are called in sequence too fast
_ <- Sync[F].sleep(cfg.containerWait.toScala)
} yield res
}
private def procPipe[F[_]](
p: String,
ctx: Context,
collect: CollectOut[F],
logger: Logger[F]
): Pipe[F, String, Unit] =
_.through(collect.append)
.map(line => s">> [${ctx.meta.nameAndVersion} ($p)] $line")
.evalMap(logger.debug(_))
/** Runs the external command that is executing the addon.
*
* If the addons specifies to collect its output, the stdout is parsed as json and
* decoded into [[AddonOutput]].
*/
def runAddonCommand[F[_]: Async](
logger: Logger[F],
cmd: SysCmd,
ctx: Context
): F[AddonResult] =
for {
stdout <-
if (ctx.meta.options.exists(_.collectOutput)) CollectOut.buffer[F]
else CollectOut.none[F].pure[F]
cmdResult <- SysExec(cmd, logger, ctx.baseDir.some)
.flatMap(
_.consumeOutputs(
procPipe("out", ctx, stdout, logger),
procPipe("err", ctx, CollectOut.none[F], logger)
)
)
.use(_.waitFor())
.attempt
addonResult <- cmdResult match {
case Right(rc) if rc != 0 =>
for {
_ <- logger.error(
s"Addon ${ctx.meta.nameAndVersion} returned non-zero: $rc"
)
} yield AddonResult.executionError(rc)
case Right(_) =>
for {
_ <- logger.debug(s"Addon ${ctx.meta.nameAndVersion} executed successfully!")
out <- stdout.get
_ <- logger.debug(s"Addon stdout: $out")
result = Option
.when(ctx.meta.options.exists(_.collectOutput) && out.nonEmpty)(
JsonParser
.decode[AddonOutput](out)
.fold(AddonResult.decodingError, AddonResult.success)
)
.getOrElse(AddonResult.empty)
} yield result
case Left(ex) =>
logger
.error(ex)(s"Executing external command failed!")
.as(AddonResult.executionFailed(ex))
}
} yield addonResult
/** Check whether `systemd-nspawn` is available on this machine. */
def checkContainer[F[_]: Async](
logger: Logger[F],
cfg: AddonExecutorConfig.NSpawn
): F[Boolean] =
for {
rc <- SysExec(cfg.nspawnVersion, logger)
.flatMap(_.logOutputs(logger, "nspawn"))
.use(_.waitFor())
_ <-
OptionT
.whenF(rc != 0)(
logger.warn(
s"No systemd-nspawn found! Addon is not executed inside a container."
)
)
.value
} yield rc == 0
}

View File

@ -0,0 +1,78 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.addons.runner
import cats.data.OptionT
import cats.effect._
import cats.kernel.Monoid
import cats.syntax.all._
import fs2.io.file.PosixPermission._
import fs2.io.file.{Files, PosixPermissions}
import docspell.addons.AddonExecutorConfig.NSpawn
import docspell.addons._
import docspell.common.Duration
import docspell.common.exec.{Args, Env, SysCmd}
import docspell.logging.Logger
final class TrivialRunner[F[_]: Async](cfg: TrivialRunner.Config) extends AddonRunner[F] {
private val sync = Async[F]
private val files = Files[F]
implicit val andMonoid: Monoid[Boolean] = Monoid.instance[Boolean](true, _ && _)
private val executeBits = PosixPermissions(
OwnerExecute,
OwnerRead,
OwnerWrite,
GroupExecute,
GroupRead,
OthersExecute,
OthersRead
)
val runnerType = List(RunnerType.Trivial)
def run(
logger: Logger[F],
env: Env,
ctx: Context
) = {
val binaryPath = ctx.meta.runner
.flatMap(_.trivial)
.map(_.exec)
.map(bin => ctx.addonDir / bin)
binaryPath match {
case None =>
sync.raiseError(new IllegalStateException("No executable specified in addon!"))
case Some(file) =>
val bin = ctx.baseDir.relativize(file)
val cmd = SysCmd(bin.toString, Args.empty).withTimeout(cfg.timeout).addEnv(env)
val withNSpawn =
OptionT
.whenF(cfg.nspawn.enabled)(RunnerUtil.checkContainer(logger, cfg.nspawn))
.getOrElse(false)
files.setPosixPermissions(file, executeBits).attempt *>
withNSpawn.flatMap {
case true =>
RunnerUtil.runInContainer(logger, cfg.nspawn, ctx)(cmd)
case false =>
RunnerUtil.runDirectly(logger, ctx)(cmd)
}
}
}
}
object TrivialRunner {
def apply[F[_]: Async](cfg: AddonExecutorConfig): TrivialRunner[F] =
new TrivialRunner[F](Config(cfg.nspawn, cfg.runTimeout))
case class Config(nspawn: NSpawn, timeout: Duration)
}

View File

@ -0,0 +1,44 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.addons
import cats.effect._
import cats.syntax.option._
import docspell.common.UrlReader
import docspell.logging.TestLoggingConfig
import munit._
class AddonArchiveTest extends CatsEffectSuite with TestLoggingConfig with Fixtures {
val logger = docspell.logging.getLogger[IO]
tempDir.test("Read archive from directory") { dir =>
for {
archive <- IO(AddonArchive(dummyAddonUrl, "", ""))
path <- archive.extractTo[IO](UrlReader.defaultReader[IO], dir)
aa <- AddonArchive.read[IO](dummyAddonUrl, UrlReader.defaultReader[IO], path.some)
_ = {
assertEquals(aa.name, "dummy-addon")
assertEquals(aa.version, "2.9")
assertEquals(aa.url, dummyAddonUrl)
}
} yield ()
}
test("Read archive from zip file") {
for {
archive <- AddonArchive.read[IO](dummyAddonUrl, UrlReader.defaultReader[IO])
_ = {
assertEquals(archive.name, "dummy-addon")
assertEquals(archive.version, "2.9")
assertEquals(archive.url, dummyAddonUrl)
}
} yield ()
}
}

View File

@ -0,0 +1,63 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.addons
import cats.effect._
import docspell.logging.{Level, TestLoggingConfig}
import munit._
class AddonExecutorTest extends CatsEffectSuite with Fixtures with TestLoggingConfig {
val logger = docspell.logging.getLogger[IO]
override def docspellLogConfig =
super.docspellLogConfig.copy(minimumLevel = Level.Trace)
tempDir.test("select docker if Dockerfile exists") { dir =>
for {
_ <- files.createFile(dir / "Dockerfile")
cfg = testExecutorConfig(
RunnerType.Docker,
RunnerType.NixFlake,
RunnerType.Trivial
)
meta = dummyAddonMeta.copy(runner = None)
r <- AddonExecutor.selectRunner[IO](cfg, meta, dir)
_ = assertEquals(r.runnerType, List(RunnerType.Docker))
} yield ()
}
tempDir.test("select nix-flake if flake.nix exists") { dir =>
for {
_ <- files.createFile(dir / "flake.nix")
cfg = testExecutorConfig(
RunnerType.Docker,
RunnerType.NixFlake,
RunnerType.Trivial
)
meta = dummyAddonMeta.copy(runner = None)
r <- AddonExecutor.selectRunner[IO](cfg, meta, dir)
_ = assertEquals(r.runnerType, List(RunnerType.NixFlake))
} yield ()
}
tempDir.test("select nix-flake and docker") { dir =>
for {
_ <- files.createFile(dir / "flake.nix")
_ <- files.createFile(dir / "Dockerfile")
cfg = testExecutorConfig(
RunnerType.Docker,
RunnerType.NixFlake,
RunnerType.Trivial
)
meta = dummyAddonMeta.copy(runner = None)
r <- AddonExecutor.selectRunner[IO](cfg, meta, dir)
_ = assertEquals(r.runnerType, List(RunnerType.Docker, RunnerType.NixFlake))
} yield ()
}
}

View File

@ -0,0 +1,37 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.addons
import cats.effect._
import docspell.common.Glob
import docspell.files.Zip
import docspell.logging.TestLoggingConfig
import munit._
class AddonMetaTest extends CatsEffectSuite with TestLoggingConfig with Fixtures {
val logger = docspell.logging.getLogger[IO]
test("read meta from zip file") {
val meta = AddonMeta.findInZip(dummyAddonUrl.readURL[IO](8192))
assertIO(meta, dummyAddonMeta)
}
tempDir.test("read meta from directory") { dir =>
for {
_ <- dummyAddonUrl
.readURL[IO](8192)
.through(Zip.unzip(8192, Glob.all))
.through(Zip.saveTo(logger, dir, moveUp = true))
.compile
.drain
meta <- AddonMeta.findInDirectory[IO](dir)
_ = assertEquals(meta, dummyAddonMeta)
} yield ()
}
}

View File

@ -0,0 +1,40 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.addons
import docspell.addons.out.AddonOutput
import io.circe.parser.decode
import munit.FunSuite
class AddonOutputTest extends FunSuite {
test("decode empty object") {
val out = decode[AddonOutput]("{}")
println(out)
}
test("decode sample output") {
val jsonStr =
"""{ "files": [
| {
| "itemId": "qZDnyGIAJsXr",
| "textFiles": {
| "HPFvIDib6eA": "HPFvIDib6eA.txt"
| },
| "pdfFiles": {
| "HPFvIDib6eA": "HPFvIDib6eA.pdf"
| }
| }
| ]
|}
|""".stripMargin
val out = decode[AddonOutput](jsonStr)
println(out)
}
}

View File

@ -0,0 +1,98 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.addons
import java.util.concurrent.atomic.AtomicInteger
import cats.effect.IO
import fs2.io.file.Path
import docspell.common.LenientUri
import docspell.common.exec.Env
import docspell.logging.{Logger, TestLoggingConfig}
import munit._
class AddonRunnerTest extends CatsEffectSuite with TestLoggingConfig {
val logger = docspell.logging.getLogger[IO]
val dummyContext = Context(
addon = AddonRef(AddonArchive(LenientUri.unsafe("http://test"), "", ""), ""),
meta = AddonMeta.empty("test", "1.0"),
baseDir = Path(""),
addonDir = Path(""),
outputDir = Path(""),
cacheDir = Path("")
)
test("firstSuccessful must stop on first success") {
val counter = new AtomicInteger(0)
val runner = new MockRunner(IO(counter.incrementAndGet()).void)
val r = AddonRunner.firstSuccessful(runner, runner, runner)
for {
_ <- r.run(logger, Env.empty, dummyContext)
_ = assertEquals(counter.get(), 1)
} yield ()
}
test("firstSuccessful must try with next on error") {
val counter = new AtomicInteger(0)
val fail = AddonRunner.failWith[IO]("failed")
val runner: AddonRunner[IO] = new MockRunner(IO(counter.incrementAndGet()).void)
val r = AddonRunner.firstSuccessful(fail, runner, runner)
for {
_ <- r.run(logger, Env.empty, dummyContext)
_ = assertEquals(counter.get(), 1)
} yield ()
}
test("do not retry on decoding errors") {
val counter = new AtomicInteger(0)
val fail = AddonRunner.pure[IO](AddonResult.decodingError("Decoding failed"))
val increment: AddonRunner[IO] = new MockRunner(IO(counter.incrementAndGet()).void)
val r = AddonRunner.firstSuccessful(fail, increment, increment)
for {
_ <- r.run(logger, Env.empty, dummyContext)
_ = assertEquals(counter.get(), 0)
} yield ()
}
test("try on errors but stop on decoding error") {
val counter = new AtomicInteger(0)
val decodeFail = AddonRunner.pure[IO](AddonResult.decodingError("Decoding failed"))
val incrementFail =
new MockRunner(IO(counter.incrementAndGet()).void)
.as(AddonResult.executionFailed(new Exception("fail")))
val increment: AddonRunner[IO] = new MockRunner(IO(counter.incrementAndGet()).void)
val r = AddonRunner.firstSuccessful(
incrementFail,
incrementFail,
decodeFail,
increment,
increment
)
for {
_ <- r.run(logger, Env.empty, dummyContext)
_ = assertEquals(counter.get(), 2)
} yield ()
}
final class MockRunner(run: IO[Unit], result: AddonResult = AddonResult.empty)
extends AddonRunner[IO] {
val runnerType = Nil
def run(
logger: Logger[IO],
env: Env,
ctx: Context
) = run.as(result)
def as(r: AddonResult) = new MockRunner(run, r)
}
}

View File

@ -0,0 +1,71 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.addons
import cats.effect._
import cats.syntax.all._
import fs2.io.file.{Files, Path, PosixPermissions}
import docspell.addons.AddonExecutorConfig._
import docspell.addons.AddonMeta._
import docspell.addons.AddonTriggerType._
import docspell.common.{Duration, LenientUri}
import docspell.logging.TestLoggingConfig
import munit.CatsEffectSuite
trait Fixtures extends TestLoggingConfig { self: CatsEffectSuite =>
val files: Files[IO] = Files[IO]
val dummyAddonUrl =
LenientUri.fromJava(getClass.getResource("/docspell-dummy-addon-master.zip"))
val dummyAddonMeta =
AddonMeta(
meta =
AddonMeta.Meta("dummy-addon", "2.9", "Some dummy addon only for testing.\n".some),
triggers = Some(
Set(Scheduled, FinalProcessItem, FinalReprocessItem)
),
None,
runner = Runner(
nix = NixRunner(true).some,
docker = DockerRunner(
enable = true,
image = None,
build = "Dockerfile".some
).some,
trivial = TrivialRunner(true, "src/addon.sh").some
).some,
options = Options(networking = true, collectOutput = true).some
)
def baseTempDir: Path =
Path(s"/tmp/target/test-temp")
val tempDir =
ResourceFixture[Path](
Resource.eval(Files[IO].createDirectories(baseTempDir)) *>
Files[IO]
.tempDirectory(baseTempDir.some, "run-", PosixPermissions.fromOctal("777"))
)
def testExecutorConfig(
runner: RunnerType,
runners: RunnerType*
): AddonExecutorConfig = {
val nspawn = NSpawn(true, "sudo", "systemd-nspawn", Duration.millis(100))
AddonExecutorConfig(
runner :: runners.toList,
Duration.minutes(2),
nspawn,
NixConfig("nix", Duration.minutes(2)),
DockerConfig("docker", Duration.minutes(2))
)
}
}

View File

@ -15,8 +15,8 @@ import fs2.io.file.{Files, Path}
import docspell.analysis.classifier
import docspell.analysis.classifier.TextClassifier._
import docspell.analysis.nlp.Properties
import docspell.common._
import docspell.common.syntax.FileSyntax._
import docspell.common.util.File
import docspell.logging.Logger
import edu.stanford.nlp.classify.ColumnDataClassifier

View File

@ -14,6 +14,7 @@ import cats.implicits._
import docspell.analysis.NlpSettings
import docspell.common._
import docspell.common.util.File
/** Creating the StanfordCoreNLP pipeline is quite expensive as it involves IO and
* initializing large objects.

View File

@ -17,6 +17,7 @@ import fs2.io.file.Files
import docspell.analysis.classifier.TextClassifier.Data
import docspell.common._
import docspell.common.util.File
import docspell.logging.TestLoggingConfig
import munit._

View File

@ -13,6 +13,7 @@ import cats.effect.unsafe.implicits.global
import docspell.analysis.Env
import docspell.common._
import docspell.common.util.File
import docspell.files.TestFiles
import docspell.logging.TestLoggingConfig

View File

@ -20,6 +20,7 @@ trait AttachedEvent[R] {
object AttachedEvent {
/** Only the result, no events. */
def only[R](v: R): AttachedEvent[R] =
new AttachedEvent[R] {
val value = v

View File

@ -8,11 +8,14 @@ package docspell.backend
import cats.effect._
import docspell.backend.BackendCommands.EventContext
import docspell.backend.auth.Login
import docspell.backend.fulltext.CreateIndex
import docspell.backend.ops._
import docspell.backend.signup.OSignup
import docspell.common.bc.BackendCommandRunner
import docspell.ftsclient.FtsClient
import docspell.joexapi.client.JoexClient
import docspell.notification.api.{EventExchange, NotificationModule}
import docspell.pubsub.api.PubSubT
import docspell.scheduler.JobStoreModule
@ -20,6 +23,7 @@ import docspell.store.Store
import docspell.totp.Totp
import emil.Emil
import org.http4s.client.Client
trait BackendApp[F[_]] {
@ -35,6 +39,7 @@ trait BackendApp[F[_]] {
def job: OJob[F]
def item: OItem[F]
def itemSearch: OItemSearch[F]
def attachment: OAttachment[F]
def fulltext: OFulltext[F]
def mail: OMail[F]
def joex: OJoex[F]
@ -52,23 +57,30 @@ trait BackendApp[F[_]] {
def fileRepository: OFileRepository[F]
def itemLink: OItemLink[F]
def downloadAll: ODownloadAll[F]
def addons: OAddons[F]
def commands(eventContext: Option[EventContext]): BackendCommandRunner[F, Unit]
}
object BackendApp {
def create[F[_]: Async](
cfg: Config,
store: Store[F],
javaEmil: Emil[F],
httpClient: Client[F],
ftsClient: FtsClient[F],
pubSubT: PubSubT[F],
schedulerModule: JobStoreModule[F],
notificationMod: NotificationModule[F]
): Resource[F, BackendApp[F]] =
for {
nodeImpl <- ONode(store)
totpImpl <- OTotp(store, Totp.default)
loginImpl <- Login[F](store, Totp.default)
signupImpl <- OSignup[F](store)
joexImpl <- OJoex(pubSubT)
joexClient = JoexClient(httpClient)
joexImpl <- OJoex(pubSubT, nodeImpl, joexClient)
collImpl <- OCollective[F](
store,
schedulerModule.userTasks,
@ -80,7 +92,6 @@ object BackendApp {
equipImpl <- OEquipment[F](store)
orgImpl <- OOrganization(store)
uploadImpl <- OUpload(store, schedulerModule.jobs)
nodeImpl <- ONode(store)
jobImpl <- OJob(store, joexImpl, pubSubT)
createIndex <- CreateIndex.resource(ftsClient, store)
itemImpl <- OItem(store, ftsClient, createIndex, schedulerModule.jobs)
@ -109,6 +120,16 @@ object BackendApp {
fileRepoImpl <- OFileRepository(store, schedulerModule.jobs)
itemLinkImpl <- Resource.pure(OItemLink(store, itemSearchImpl))
downloadAllImpl <- Resource.pure(ODownloadAll(store, jobImpl, schedulerModule.jobs))
attachImpl <- Resource.pure(OAttachment(store, ftsClient, schedulerModule.jobs))
addonsImpl <- Resource.pure(
OAddons(
cfg.addons,
store,
schedulerModule.userTasks,
schedulerModule.jobs,
joexImpl
)
)
} yield new BackendApp[F] {
val pubSub = pubSubT
val login = loginImpl
@ -139,5 +160,10 @@ object BackendApp {
val fileRepository = fileRepoImpl
val itemLink = itemLinkImpl
val downloadAll = downloadAllImpl
val addons = addonsImpl
val attachment = attachImpl
def commands(eventContext: Option[EventContext]) =
BackendCommands.fromBackend(this, eventContext)
}
}

View File

@ -0,0 +1,175 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.backend
import cats.data.{NonEmptyList => Nel}
import cats.effect.Sync
import cats.syntax.all._
import docspell.backend.BackendCommands.EventContext
import docspell.backend.ops.OCustomFields.SetValue
import docspell.backend.ops._
import docspell.common.bc._
import docspell.common.{AccountId, Ident, LenientUri}
private[backend] class BackendCommands[F[_]: Sync](
itemOps: OItem[F],
attachOps: OAttachment[F],
fieldOps: OCustomFields[F],
notificationOps: ONotification[F],
eventContext: Option[EventContext]
) extends BackendCommandRunner[F, Unit] {
private[this] val logger = docspell.logging.getLogger[F]
def run(collective: Ident, cmd: BackendCommand): F[Unit] =
doRun(collective, cmd).attempt.flatMap {
case Right(_) => ().pure[F]
case Left(ex) =>
logger.error(ex)(s"Backend command $cmd failed for collective ${collective.id}.")
}
def doRun(collective: Ident, cmd: BackendCommand): F[Unit] =
cmd match {
case BackendCommand.ItemUpdate(item, actions) =>
actions.traverse_(a => runItemAction(collective, item, a))
case BackendCommand.AttachmentUpdate(item, attach, actions) =>
actions.traverse_(a => runAttachAction(collective, item, attach, a))
}
def runAll(collective: Ident, cmds: List[BackendCommand]): F[Unit] =
cmds.traverse_(run(collective, _))
def runItemAction(collective: Ident, item: Ident, action: ItemAction): F[Unit] =
action match {
case ItemAction.AddTags(tags) =>
logger.debug(s"Setting tags $tags on ${item.id} for ${collective.id}") *>
itemOps
.linkTags(item, tags.toList, collective)
.flatMap(sendEvents)
case ItemAction.RemoveTags(tags) =>
logger.debug(s"Remove tags $tags on ${item.id} for ${collective.id}") *>
itemOps
.removeTagsMultipleItems(Nel.of(item), tags.toList, collective)
.flatMap(sendEvents)
case ItemAction.ReplaceTags(tags) =>
logger.debug(s"Replace tags $tags on ${item.id} for ${collective.id}") *>
itemOps
.setTags(item, tags.toList, collective)
.flatMap(sendEvents)
case ItemAction.SetFolder(folder) =>
logger.debug(s"Set folder $folder on ${item.id} for ${collective.id}") *>
itemOps
.setFolder(item, folder, collective)
.void
case ItemAction.RemoveTagsCategory(cats) =>
logger.debug(
s"Remove tags in categories $cats on ${item.id} for ${collective.id}"
) *>
itemOps
.removeTagsOfCategories(item, collective, cats)
.flatMap(sendEvents)
case ItemAction.SetCorrOrg(id) =>
logger.debug(
s"Set correspondent organization ${id.map(_.id)} for ${collective.id}"
) *>
itemOps.setCorrOrg(Nel.of(item), id, collective).void
case ItemAction.SetCorrPerson(id) =>
logger.debug(
s"Set correspondent person ${id.map(_.id)} for ${collective.id}"
) *>
itemOps.setCorrPerson(Nel.of(item), id, collective).void
case ItemAction.SetConcPerson(id) =>
logger.debug(
s"Set concerning person ${id.map(_.id)} for ${collective.id}"
) *>
itemOps.setConcPerson(Nel.of(item), id, collective).void
case ItemAction.SetConcEquipment(id) =>
logger.debug(
s"Set concerning equipment ${id.map(_.id)} for ${collective.id}"
) *>
itemOps.setConcEquip(Nel.of(item), id, collective).void
case ItemAction.SetField(field, value) =>
logger.debug(
s"Set field on item ${item.id} ${field.id} to '$value' for ${collective.id}"
) *>
fieldOps
.setValue(item, SetValue(field, value, collective))
.flatMap(sendEvents)
case ItemAction.SetNotes(notes) =>
logger.debug(s"Set notes on item ${item.id} for ${collective.id}") *>
itemOps.setNotes(item, notes, collective).void
case ItemAction.AddNotes(notes, sep) =>
logger.debug(s"Add notes on item ${item.id} for ${collective.id}") *>
itemOps.addNotes(item, notes, sep, collective).void
case ItemAction.SetName(name) =>
logger.debug(s"Set name '$name' on item ${item.id} for ${collective.id}") *>
itemOps.setName(item, name, collective).void
}
def runAttachAction(
collective: Ident,
itemId: Ident,
attachId: Ident,
action: AttachmentAction
): F[Unit] =
action match {
case AttachmentAction.SetExtractedText(text) =>
attachOps.setExtractedText(
collective,
itemId,
attachId,
text.getOrElse("").pure[F]
)
}
private def sendEvents(result: AttachedEvent[_]): F[Unit] =
eventContext match {
case Some(ctx) =>
notificationOps.offerEvents(result.event(ctx.account, ctx.baseUrl))
case None => ().pure[F]
}
}
object BackendCommands {
/** If supplied, notification events will be send. */
case class EventContext(account: AccountId, baseUrl: Option[LenientUri])
def fromBackend[F[_]: Sync](
backendApp: BackendApp[F],
eventContext: Option[EventContext] = None
): BackendCommandRunner[F, Unit] =
new BackendCommands[F](
backendApp.item,
backendApp.attachment,
backendApp.customFields,
backendApp.notification,
eventContext
)
def apply[F[_]: Sync](
item: OItem[F],
attachment: OAttachment[F],
fields: OCustomFields[F],
notification: ONotification[F],
eventContext: Option[EventContext] = None
): BackendCommandRunner[F, Unit] =
new BackendCommands[F](item, attachment, fields, notification, eventContext)
}

View File

@ -20,7 +20,8 @@ case class Config(
mailDebug: Boolean,
jdbc: JdbcConfig,
signup: SignupConfig,
files: Config.Files
files: Config.Files,
addons: Config.Addons
) {
def mailSettings: Settings =
@ -66,4 +67,21 @@ object Config {
(storesEmpty |+| defaultStorePresent).map(_ => this)
}
}
case class Addons(
enabled: Boolean,
allowImpure: Boolean,
allowedUrls: UrlMatcher,
deniedUrls: UrlMatcher
) {
def isAllowed(url: LenientUri): Boolean =
allowedUrls.matches(url) && !deniedUrls.matches(url)
def isDenied(url: LenientUri): Boolean =
!isAllowed(url)
}
object Addons {
val disabled: Addons =
Addons(false, false, UrlMatcher.False, UrlMatcher.True)
}
}

View File

@ -16,6 +16,26 @@ import docspell.notification.api.PeriodicQueryArgs
import docspell.scheduler.Job
object JobFactory extends MailAddressCodec {
def existingItemAddon[F[_]: Sync](
args: ItemAddonTaskArgs,
submitter: AccountId
): F[Job[ItemAddonTaskArgs]] =
Job.createNew(
ItemAddonTaskArgs.taskName,
submitter.collective,
args,
"Run addons on item",
submitter.user,
Priority.High,
args.addonRunConfigs
.map(_.take(23))
.toList
.sorted
.foldLeft(args.itemId)(_ / _)
.take(250)
.some
)
def downloadZip[F[_]: Sync](
args: DownloadZipArgs,
summaryId: Ident,

View File

@ -45,7 +45,14 @@ object CreateIndex {
chunkSize: Int
): F[Unit] = {
val attachs = store
.transact(QAttachment.allAttachmentMetaAndName(collective, itemIds, chunkSize))
.transact(
QAttachment.allAttachmentMetaAndName(
collective,
itemIds,
ItemState.validStates,
chunkSize
)
)
.map(caa =>
TextData
.attachment(

View File

@ -0,0 +1,17 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.backend.joex
import fs2.io.file.Path
import docspell.addons.AddonExecutorConfig
final case class AddonEnvConfig(
workingDir: Path,
cacheDir: Path,
executorConfig: AddonExecutorConfig
)

View File

@ -0,0 +1,199 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.backend.joex
import cats.data.OptionT
import cats.effect._
import cats.syntax.all._
import docspell.addons._
import docspell.backend.joex.AddonOps.{AddonRunConfigRef, ExecResult}
import docspell.backend.ops.OAttachment
import docspell.common._
import docspell.common.bc.BackendCommandRunner
import docspell.common.exec.Env
import docspell.logging.Logger
import docspell.scheduler.JobStore
import docspell.store.Store
import docspell.store.file.FileUrlReader
import docspell.store.records.AddonRunConfigResolved
trait AddonOps[F[_]] {
def execAll(
collective: Ident,
trigger: Set[AddonTriggerType],
runConfigIds: Set[Ident],
logger: Option[Logger[F]]
)(
middleware: Middleware[F]
): F[ExecResult]
def execById(collective: Ident, runConfigId: Ident, logger: Logger[F])(
middleware: Middleware[F]
): F[ExecResult]
/** Find enabled addon run config references to be executed. Can be additionally
* filtered by given ids and triggers.
*/
def findAddonRefs(
collective: Ident,
trigger: Set[AddonTriggerType],
runConfigIds: Set[Ident]
): F[List[AddonRunConfigRef]]
/** Find enabled addon run config reference given an addon task id */
def findAddonRef(collective: Ident, runConfigId: Ident): F[Option[AddonRunConfigRef]]
/** Creates an executor for addons given a configuration. */
def getExecutor(cfg: AddonExecutorConfig): F[AddonExecutor[F]]
}
object AddonOps {
case class AddonRunConfigRef(
id: Ident,
collective: Ident,
userId: Option[Ident],
name: String,
refs: List[AddonRef]
)
object AddonRunConfigRef {
def fromResolved(r: AddonRunConfigResolved): AddonRunConfigRef =
AddonRunConfigRef(
r.config.id,
r.config.cid,
r.config.userId,
r.config.name,
r.refs.map(ref => AddonRef(ref.archive.asArchive, ref.ref.args))
)
}
case class ExecResult(
result: List[AddonExecutionResult],
runConfigs: List[AddonRunConfigRef]
) {
lazy val combined = result.combineAll
}
object ExecResult {
def runConfigNotFound(id: Ident): ExecResult =
ExecResult(
AddonExecutionResult(
AddonResult.executionFailed(
new Exception(s"Addon run config ${id.id} not found.")
) :: Nil,
false
) :: Nil,
Nil
)
}
def apply[F[_]: Async](
cfg: AddonEnvConfig,
store: Store[F],
cmdRunner: BackendCommandRunner[F, Unit],
attachment: OAttachment[F],
jobStore: JobStore[F]
): AddonOps[F] =
new AddonOps[F] with LoggerExtension {
private[this] val logger = docspell.logging.getLogger[F]
private val urlReader = FileUrlReader(store.fileRepo)
private val postProcess = AddonPostProcess(cmdRunner, store, attachment, jobStore)
private val prepare = new AddonPrepare[F](store)
def execAll(
collective: Ident,
trigger: Set[AddonTriggerType],
runConfigIds: Set[Ident],
logger: Option[Logger[F]]
)(
custom: Middleware[F]
): F[ExecResult] =
for {
runCfgs <- findAddonRefs(collective, trigger, runConfigIds)
log = logger.getOrElse(this.logger)
_ <- log.info(s"Running ${runCfgs.size} addon tasks for trigger $trigger")
results <- runCfgs.traverse(r => execRunConfig(log, r, custom))
} yield ExecResult(results.flatMap(_.result), runCfgs)
def execById(collective: Ident, runConfigId: Ident, logger: Logger[F])(
custom: Middleware[F]
): F[ExecResult] =
(for {
runCfg <- OptionT(findAddonRef(collective, runConfigId))
execRes <- OptionT.liftF(execRunConfig(logger, runCfg, custom))
} yield execRes).getOrElse(ExecResult.runConfigNotFound(runConfigId))
def execRunConfig(
logger: Logger[F],
runCfg: AddonRunConfigRef,
custom: Middleware[F]
): F[ExecResult] =
for {
executor <- getExecutor(cfg.executorConfig)
log = logger.withRunConfig(runCfg)
result <-
Directory.temp(cfg.workingDir, "addon-output-").use { outDir =>
val cacheDir = cfg.cacheDir / runCfg.id.id
val inputEnv =
InputEnv(runCfg.refs, cfg.workingDir, outDir, cacheDir, Env.empty)
for {
middleware <- createMiddleware(custom, runCfg)
res <- middleware(executor.execute(log)).run(inputEnv)
_ <- log.debug(s"Addon result: $res")
_ <- postProcess.onResult(log, runCfg.collective, res, outDir)
} yield res
}
execRes = ExecResult(List(result), List(runCfg))
} yield execRes
def createMiddleware(custom: Middleware[F], runCfg: AddonRunConfigRef) = for {
dscMW <- prepare.createDscEnv(runCfg, cfg.executorConfig.runTimeout)
mm = dscMW >> custom >> prepare.logResult(logger, runCfg) >> Middleware
.ephemeralRun[F]
} yield mm
def getExecutor(cfg: AddonExecutorConfig): F[AddonExecutor[F]] =
Async[F].pure(AddonExecutor(cfg, urlReader))
def findAddonRefs(
collective: Ident,
trigger: Set[AddonTriggerType],
runConfigIds: Set[Ident]
): F[List[AddonRunConfigRef]] =
store
.transact(
AddonRunConfigResolved.findAllForCollective(
collective,
enabled = true.some,
trigger,
runConfigIds
)
)
.map(_.map(AddonRunConfigRef.fromResolved))
def findAddonRef(
collective: Ident,
runConfigId: Ident
): F[Option[AddonRunConfigRef]] =
OptionT(
store
.transact(
AddonRunConfigResolved.findById(
runConfigId,
collective,
enabled = Some(true)
)
)
).map(AddonRunConfigRef.fromResolved).value
}
}

View File

@ -0,0 +1,198 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.backend.joex
import cats.data.OptionT
import cats.effect.kernel.Sync
import cats.syntax.all._
import fs2.io.file.{Files, Path}
import docspell.addons._
import docspell.addons.out.{AddonOutput, ItemFile, NewItem}
import docspell.backend.JobFactory
import docspell.backend.ops.OAttachment
import docspell.common._
import docspell.common.bc.BackendCommandRunner
import docspell.files.FileSupport
import docspell.logging.Logger
import docspell.scheduler.JobStore
import docspell.store.Store
import docspell.store.records._
final private[joex] class AddonPostProcess[F[_]: Sync: Files](
cmdRunner: BackendCommandRunner[F, Unit],
store: Store[F],
attachOps: OAttachment[F],
jobStore: JobStore[F]
) extends FileSupport {
def onResult(
logger: Logger[F],
collective: Ident,
result: AddonExecutionResult,
outputDir: Path
): F[Unit] =
result.addonResult match {
case AddonResult.Success(output) =>
onSuccess(logger, collective, output, outputDir)
case _ =>
().pure[F]
}
def onSuccess(
logger: Logger[F],
collective: Ident,
output: AddonOutput,
outputDir: Path
): F[Unit] =
for {
_ <- logger.info("Applying addon output")
_ <- cmdRunner.runAll(collective, output.commands)
_ <- logger.debug("Applying changes from files")
_ <- output.files.traverse_(updateOne(logger, collective, outputDir))
_ <- output.newItems.traverse_(submitNewItem(logger, collective, outputDir))
} yield ()
def submitNewItem(
logger: Logger[F],
collective: Ident,
outputDir: Path
)(newItem: NewItem): F[Unit] =
for {
_ <- logger.info(s"Submit new item with ${newItem.files.size} files")
files <- newItem.resolveFiles[F](logger, outputDir)
collLang <- store.transact(RCollective.findLanguage(collective))
uploaded <- files.traverse(file =>
file.readAll
.through(
store.fileRepo.save(
collective,
FileCategory.AttachmentSource,
MimeTypeHint.filename(file)
)
)
.compile
.lastOrError
.map(key => file.fileName.toString -> key)
)
_ <- logger.debug(s"Saved ${uploaded.size} files to be processed.")
args = ProcessItemArgs(
newItem.toProcessMeta(collective, collLang, "addon"),
uploaded.map(f => ProcessItemArgs.File(f._1.some, f._2))
)
account = AccountId(collective, DocspellSystem.user)
job <- JobFactory.processItem(args, account, Priority.High, None)
_ <- jobStore.insert(job.encode)
_ <- logger.debug(s"Submitted job for processing: ${job.id}")
} yield ()
def updateOne(logger: Logger[F], collective: Ident, outputDir: Path)(
itemFile: ItemFile
): F[Unit] =
for {
textFiles <- itemFile.resolveTextFiles(logger, outputDir)
pdfFiles <- itemFile.resolvePdfFiles(logger, outputDir)
previewFiles <- itemFile.resolvePreviewFiles(logger, outputDir)
attachs <- OptionT
.whenF(textFiles.nonEmpty || pdfFiles.nonEmpty || previewFiles.nonEmpty)(
store.transact(RAttachment.findByItem(itemFile.itemId))
)
.getOrElse(Vector.empty)
_ <- textFiles.traverse_ { case (key, file) =>
withAttach(logger, key, attachs) { ra =>
setText(collective, ra, file.readText)
}
}
_ <- pdfFiles.traverse_ { case (key, file) =>
withAttach(logger, key, attachs) { ra =>
replacePdf(collective, ra, file, previewFiles.forall(_._1 != key))
}
}
_ <- previewFiles.traverse_ { case (key, file) =>
withAttach(logger, key, attachs) { ra =>
replacePreview(collective, ra.id, file)
}
}
_ <- submitNewFiles(logger, collective, outputDir)(itemFile)
} yield ()
def submitNewFiles(
logger: Logger[F],
collective: Ident,
outputDir: Path
)(itemFile: ItemFile): F[Unit] =
for {
_ <- logger.info(s"Submitting new file for item")
collLang <- store.transact(RCollective.findLanguage(collective))
newFiles <- itemFile.resolveNewFiles(logger, outputDir)
byMeta = newFiles.groupBy(_._1.metadata).view.mapValues(_.map(_._2))
account = AccountId(collective, DocspellSystem.user)
_ <- byMeta.toList.traverse_ { case (meta, files) =>
for {
uploaded <- files.traverse(file =>
file.readAll
.through(
store.fileRepo.save(
collective,
FileCategory.AttachmentSource,
MimeTypeHint.filename(file)
)
)
.compile
.lastOrError
.map(key => file.fileName.toString -> key)
)
args = ProcessItemArgs(
meta.toProcessMeta(collective, itemFile.itemId, collLang, "addon"),
uploaded.map(f => ProcessItemArgs.File(f._1.some, f._2))
)
job <- JobFactory.processItem(args, account, Priority.High, None)
_ <- jobStore.insert(job.encode)
_ <- logger.debug(s"Submitted job for processing: ${job.id}")
} yield ()
}
} yield ()
private def withAttach(logger: Logger[F], key: String, attachs: Vector[RAttachment])(
run: RAttachment => F[Unit]
): F[Unit] =
OptionT
.fromOption(
attachs.find(a => a.id.id == key || key.toIntOption == a.position.some)
)
.semiflatMap(run)
.getOrElseF(logger.warn(s"Cannot find attachment for $key to update text!"))
private def setText(collective: Ident, ra: RAttachment, readText: F[String]): F[Unit] =
attachOps.setExtractedText(collective, ra.itemId, ra.id, readText)
private def replacePdf(
collective: Ident,
ra: RAttachment,
file: Path,
generatePreview: Boolean
): F[Unit] =
attachOps.addOrReplacePdf(collective, ra.id, file.readAll, generatePreview)
private def replacePreview(
collective: Ident,
attachId: Ident,
imageData: Path
): F[Unit] =
attachOps.addOrReplacePreview(collective, attachId, imageData.readAll)
}
object AddonPostProcess {
def apply[F[_]: Sync: Files](
cmdRunner: BackendCommandRunner[F, Unit],
store: Store[F],
attachment: OAttachment[F],
jobStore: JobStore[F]
): AddonPostProcess[F] =
new AddonPostProcess[F](cmdRunner, store, attachment, jobStore)
}

View File

@ -0,0 +1,75 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.backend.joex
import cats.data.{Kleisli, OptionT}
import cats.effect._
import cats.syntax.all._
import docspell.addons.Middleware
import docspell.backend.auth.AuthToken
import docspell.backend.joex.AddonOps.AddonRunConfigRef
import docspell.common._
import docspell.logging.Logger
import docspell.store.Store
import docspell.store.records.{RNode, RUser}
import scodec.bits.ByteVector
private[joex] class AddonPrepare[F[_]: Sync](store: Store[F]) extends LoggerExtension {
def logResult(logger: Logger[F], ref: AddonRunConfigRef): Middleware[F] =
Middleware(_.mapF(_.attempt.flatTap {
case Right(_) => ().pure[F]
case Left(ex) =>
logger
.withRunConfig(ref)
.warn(ex)(s"Addon task '${ref.id.id}' has failed")
}.rethrow))
/** Creates environment variables for dsc to connect to the docspell server for the
* given run config.
*/
def createDscEnv(
runConfigRef: AddonRunConfigRef,
tokenValidity: Duration
): F[Middleware[F]] =
(for {
userId <- OptionT.fromOption[F](runConfigRef.userId)
user <- OptionT(store.transact(RUser.getIdByIdOrLogin(userId)))
account = AccountId(runConfigRef.collective, user.login)
env =
Middleware.prepare[F](
Kleisli(input => makeDscEnv(account, tokenValidity).map(input.addEnv))
)
} yield env).getOrElse(Middleware.identity[F])
/** Creates environment variables to have dsc automatically connect as the given user.
* Additionally a random rest-server is looked up from the database to set its url.
*/
def makeDscEnv(
accountId: AccountId,
tokenValidity: Duration
): F[Map[String, String]] =
for {
serverNode <- store.transact(
RNode
.findAll(NodeType.Restserver)
.map(_.sortBy(_.updated).lastOption)
)
url = serverNode.map(_.url).map(u => "DSC_DOCSPELL_URL" -> u.asString)
secret = serverNode.flatMap(_.serverSecret)
token <- AuthToken.user(
accountId,
false,
secret.getOrElse(ByteVector.empty),
tokenValidity.some
)
session = ("DSC_SESSION" -> token.asString).some
} yield List(url, session).flatten.toMap
}

View File

@ -0,0 +1,18 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.backend.joex
import docspell.backend.joex.AddonOps.AddonRunConfigRef
import docspell.logging.Logger
trait LoggerExtension {
implicit final class LoggerDataOps[F[_]](self: Logger[F]) {
def withRunConfig(t: AddonRunConfigRef): Logger[F] =
self.capture("addon-task-id", t.id)
}
}

View File

@ -0,0 +1,48 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.backend.ops
import cats.data.NonEmptyList
import docspell.addons.{AddonArchive, AddonMeta, AddonTriggerType}
sealed trait AddonRunConfigError {
final def cast: AddonRunConfigError = this
def toLeft[A]: Either[AddonRunConfigError, A] = Left(this)
def message: String
}
object AddonRunConfigError {
case object MissingSchedule extends AddonRunConfigError {
val message =
"The run config has a trigger 'scheduled' but doesn't provide a schedule!"
}
case object ObsoleteSchedule extends AddonRunConfigError {
val message = "The run config has a schedule, but not a trigger 'Scheduled'."
}
case class MismatchingTrigger(unsupported: NonEmptyList[(String, AddonTriggerType)])
extends AddonRunConfigError {
def message: String = {
val list =
unsupported.map { case (name, tt) => s"$name: ${tt.name}" }.toList.mkString(", ")
s"Some listed addons don't support all defined triggers: $list"
}
}
object MismatchingTrigger {
def apply(addon: AddonMeta, tt: AddonTriggerType): MismatchingTrigger =
MismatchingTrigger(NonEmptyList.of(addon.nameAndVersion -> tt))
def apply(addon: AddonArchive, tt: AddonTriggerType): MismatchingTrigger =
MismatchingTrigger(NonEmptyList.of(addon.nameAndVersion -> tt))
}
}

View File

@ -0,0 +1,54 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.backend.ops
import cats.data.NonEmptyList
import cats.effect._
import cats.syntax.all._
import docspell.backend.ops.AddonRunConfigError._
import docspell.backend.ops.OAddons.{AddonRunConfigResult, AddonRunInsert}
import docspell.common.Ident
import docspell.store.Store
import docspell.store.records.RAddonArchive
object AddonRunConfigValidate {
def apply[F[_]: Sync](store: Store[F], cid: Ident)(
cfg: AddonRunInsert
): F[AddonRunConfigResult[AddonRunInsert]] = {
val init: AddonRunConfigResult[Unit] = ().asRight
List(
checkScheduled(cfg).pure[F],
checkTriggers(store, cid)(cfg)
)
.foldLeftM(init)((res, fr) => fr.map(r => res.flatMap(_ => r)))
.map(_.as(cfg))
}
def checkTriggers[F[_]: Sync](store: Store[F], cid: Ident)(
cfg: AddonRunInsert
): F[AddonRunConfigResult[Unit]] =
for {
addons <- store.transact(RAddonArchive.findByIds(cid, cfg.addons.map(_.addonId)))
given = cfg.triggered.toList.toSet
res = addons
.flatMap(r => given.diff(r.triggers).map(tt => r.nameAndVersion -> tt))
maybeError = NonEmptyList
.fromList(res)
.map(nel => MismatchingTrigger(nel))
} yield maybeError.map(_.toLeft).getOrElse(Right(()))
def checkScheduled(cfg: AddonRunInsert): AddonRunConfigResult[Unit] =
(cfg.isScheduled, cfg.schedule) match {
case (true, None) => MissingSchedule.toLeft[Unit]
case (false, Some(_)) => ObsoleteSchedule.toLeft[Unit]
case _ => ().asRight
}
}

View File

@ -0,0 +1,156 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.backend.ops
import cats.data.EitherT
import cats.effect._
import cats.syntax.all._
import fs2.Stream
import fs2.io.file.Path
import docspell.addons.{AddonMeta, RunnerType}
import docspell.backend.Config
import docspell.backend.ops.AddonValidationError._
import docspell.backend.ops.OAddons.AddonValidationResult
import docspell.common.{Ident, LenientUri, UrlReader}
import docspell.joexapi.model.AddonSupport
import docspell.store.Store
import docspell.store.records.RAddonArchive
final class AddonValidate[F[_]: Async](
cfg: Config.Addons,
store: Store[F],
joexOps: OJoex[F]
) {
private[this] val logger = docspell.logging.getLogger[F]
def fromUrl(
collective: Ident,
url: LenientUri,
reader: UrlReader[F],
localUrl: Option[LenientUri] = None,
checkExisting: Boolean = true
): F[AddonValidationResult[AddonMeta]] =
if (!cfg.enabled) AddonsDisabled.resultF
else if (cfg.isDenied(url)) UrlUntrusted(url).resultF
else if (checkExisting)
store.transact(RAddonArchive.findByUrl(collective, url)).flatMap {
case Some(ar) =>
AddonExists("An addon with this url already exists!", ar).resultF
case None =>
archive(collective, reader(localUrl.getOrElse(url)).asRight, checkExisting)
}
else archive(collective, reader(localUrl.getOrElse(url)).asRight, checkExisting)
def archive(
collective: Ident,
addonData: Either[Path, Stream[F, Byte]],
checkExisting: Boolean = true
): F[AddonValidationResult[AddonMeta]] =
(for {
_ <- EitherT.cond[F](cfg.enabled, (), AddonsDisabled.cast)
meta <-
EitherT(
addonData
.fold(
AddonMeta.findInDirectory[F],
AddonMeta.findInZip[F]
)
.attempt
)
.leftMap(ex => NotAnAddon(ex).cast)
_ <- EitherT.cond(
meta.triggers.exists(_.nonEmpty),
(),
InvalidAddon(
"The addon doesn't define any triggers. At least one is required!"
).cast
)
_ <- EitherT.cond(
meta.options.exists(_.isUseful),
(),
InvalidAddon(
"Addon defines no output and no networking. It can't do anything useful."
).cast
)
_ <- EitherT.cond(cfg.allowImpure || meta.isPure, (), ImpureAddonsDisabled.cast)
_ <-
if (checkExisting)
EitherT(
store
.transact(
RAddonArchive
.findByNameAndVersion(collective, meta.meta.name, meta.meta.version)
)
.map {
case Some(ar) => AddonExists(ar).result
case None => rightUnit
}
)
else rightUnitT
joexSupport <- EitherT.liftF(joexOps.getAddonSupport)
addonRunners <- EitherT.liftF(meta.enabledTypes(addonData))
_ <- EitherT.liftF(
logger.info(
s"Comparing joex support vs addon runner: $joexSupport vs. $addonRunners"
)
)
_ <- EitherT.fromEither(validateJoexSupport(addonRunners, joexSupport))
} yield meta).value
private def validateJoexSupport(
addonRunnerTypes: List[RunnerType],
joexSupport: List[AddonSupport]
): AddonValidationResult[Unit] = {
val addonRunners = addonRunnerTypes.mkString(", ")
for {
_ <- Either.cond(
joexSupport.nonEmpty,
(),
AddonUnsupported("There are no joex nodes that have addons enabled!", Nil).cast
)
_ <- Either.cond(
addonRunners.nonEmpty,
(),
InvalidAddon("The addon doesn't enable any runner.")
)
ids = joexSupport
.map(n => n.nodeId -> n.runners.intersect(addonRunnerTypes).toSet)
unsupportedJoex = ids.filter(_._2.isEmpty).map(_._1)
_ <- Either.cond(
ids.forall(_._2.nonEmpty),
(),
AddonUnsupported(
s"A joex node doesn't support this addons runners: $addonRunners. " +
s"Check: ${unsupportedJoex.map(_.id).mkString(", ")}.",
unsupportedJoex
).cast
)
} yield ()
}
private def rightUnit: AddonValidationResult[Unit] =
().asRight[AddonValidationError]
private def rightUnitT: EitherT[F, AddonValidationError, Unit] =
EitherT.fromEither(rightUnit)
implicit final class ErrorOps(self: AddonValidationError) {
def result: AddonValidationResult[AddonMeta] =
self.toLeft
def resultF: F[AddonValidationResult[AddonMeta]] =
result.pure[F]
}
}

View File

@ -0,0 +1,85 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.backend.ops
import docspell.common.{Ident, LenientUri}
import docspell.store.records.RAddonArchive
import io.circe.generic.extras.Configuration
import io.circe.generic.extras.semiauto.{deriveConfiguredDecoder, deriveConfiguredEncoder}
import io.circe.generic.semiauto.{deriveDecoder, deriveEncoder}
import io.circe.{Decoder, Encoder}
sealed trait AddonValidationError {
def cast: AddonValidationError = this
def toLeft[A]: Either[AddonValidationError, A] = Left(this)
}
object AddonValidationError {
implicit private val throwableDecoder: Decoder[Throwable] =
Decoder.decodeString.map(new Exception(_))
implicit private val throwableEncoder: Encoder[Throwable] =
Encoder.encodeString.contramap(_.getMessage)
case object AddonsDisabled extends AddonValidationError {}
case class UrlUntrusted(url: LenientUri) extends AddonValidationError
object UrlUntrusted {
implicit val jsonDecoder: Decoder[UrlUntrusted] = deriveDecoder
implicit val jsonEncoder: Encoder[UrlUntrusted] = deriveEncoder
}
case class NotAnAddon(error: Throwable) extends AddonValidationError
object NotAnAddon {
implicit val jsonDecoder: Decoder[NotAnAddon] = deriveDecoder
implicit val jsonEncoder: Encoder[NotAnAddon] = deriveEncoder
}
case class AddonUnsupported(message: String, affectedNodes: List[Ident])
extends AddonValidationError
object AddonUnsupported {
implicit val jsonDecoder: Decoder[AddonUnsupported] = deriveDecoder
implicit val jsonEncoder: Encoder[AddonUnsupported] = deriveEncoder
}
case class InvalidAddon(message: String) extends AddonValidationError
object InvalidAddon {
implicit val jsonDecoder: Decoder[InvalidAddon] = deriveDecoder
implicit val jsonEncoder: Encoder[InvalidAddon] = deriveEncoder
}
case class AddonExists(message: String, addon: RAddonArchive)
extends AddonValidationError
object AddonExists {
def apply(addon: RAddonArchive): AddonExists =
AddonExists(s"An addon '${addon.name}/${addon.version}' already exists!", addon)
implicit val jsonDecoder: Decoder[AddonExists] = deriveDecoder
implicit val jsonEncoder: Encoder[AddonExists] = deriveEncoder
}
case object AddonNotFound extends AddonValidationError
case class DownloadFailed(error: Throwable) extends AddonValidationError
object DownloadFailed {
implicit val jsonDecoder: Decoder[DownloadFailed] = deriveDecoder
implicit val jsonEncoder: Encoder[DownloadFailed] = deriveEncoder
}
case object ImpureAddonsDisabled extends AddonValidationError
case object RefreshLocalAddon extends AddonValidationError
implicit val jsonConfig: Configuration =
Configuration.default.withKebabCaseConstructorNames
.withDiscriminator("errorType")
implicit val jsonDecoder: Decoder[AddonValidationError] = deriveConfiguredDecoder
implicit val jsonEncoder: Encoder[AddonValidationError] = deriveConfiguredEncoder
}

View File

@ -0,0 +1,426 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.backend.ops
import cats.data.{EitherT, NonEmptyList, OptionT}
import cats.effect._
import cats.syntax.all._
import docspell.addons.{AddonMeta, AddonTriggerType}
import docspell.backend.ops.AddonValidationError._
import docspell.backend.ops.OAddons._
import docspell.backend.{Config, JobFactory}
import docspell.common._
import docspell.logging.Logger
import docspell.scheduler.JobStore
import docspell.scheduler.usertask.{UserTask, UserTaskScope, UserTaskStore}
import docspell.store.Store
import docspell.store.file.FileUrlReader
import docspell.store.records._
import com.github.eikek.calev.CalEvent
trait OAddons[F[_]] {
/** Registers a new addon. An error is returned if an addon with this url already
* exists.
*/
def registerAddon(
collective: Ident,
url: LenientUri,
logger: Option[Logger[F]]
): F[AddonValidationResult[(RAddonArchive, AddonMeta)]]
/** Refreshes an existing addon by downloading it again and updating metadata. */
def refreshAddon(
collective: Ident,
addonId: Ident
): F[AddonValidationResult[(RAddonArchive, AddonMeta)]]
/** Look into the addon at the given url and return its metadata. */
def inspectAddon(
collective: Ident,
url: LenientUri
): F[AddonValidationResult[AddonMeta]]
/** Deletes the addon if it exists. */
def deleteAddon(collective: Ident, addonId: Ident): F[Boolean]
def getAllAddons(collective: Ident): F[List[RAddonArchive]]
/** Inserts or updates the addon run configuration. If it already exists (and the given
* id is non empty), it will be completely replaced with the given one.
*/
def upsertAddonRunConfig(
collective: Ident,
runConfig: AddonRunInsert
): F[AddonRunConfigResult[Ident]]
/** Deletes this task from the database. */
def deleteAddonRunConfig(collective: Ident, runConfigId: Ident): F[Boolean]
def getAllAddonRunConfigs(collective: Ident): F[List[AddonRunInfo]]
def runAddonForItem(
account: AccountId,
itemIds: NonEmptyList[Ident],
addonRunConfigIds: Set[Ident]
): F[Unit]
}
object OAddons {
val scheduledAddonTaskName: Ident =
ScheduledAddonTaskArgs.taskName
case class AddonRunInsert(
id: Ident,
name: String,
enabled: Boolean,
userId: Option[Ident],
schedule: Option[CalEvent],
triggered: NonEmptyList[AddonTriggerType],
addons: NonEmptyList[AddonArgs]
) {
def isScheduled: Boolean =
triggered.exists(_ == AddonTriggerType.Scheduled)
}
case class AddonArgs(addonId: Ident, args: String)
case class AddonRunInfo(
id: Ident,
name: String,
enabled: Boolean,
userId: Option[Ident],
schedule: Option[CalEvent],
triggered: List[AddonTriggerType],
addons: List[(RAddonArchive, RAddonRunConfigAddon)]
)
object AddonRunInfo {
def fromRunConfigData(
timer: Option[CalEvent],
addons: List[(RAddonArchive, RAddonRunConfigAddon)]
)(t: AddonRunConfigData): AddonRunInfo =
AddonRunInfo(
id = t.runConfig.id,
name = t.runConfig.name,
enabled = t.runConfig.enabled,
userId = t.runConfig.userId,
schedule = timer,
triggered = t.triggers.map(_.trigger),
addons = addons
)
}
type AddonRunConfigResult[A] = Either[AddonRunConfigError, A]
object AddonRunConfigResult {
def success[A](value: A): AddonRunConfigResult[A] = Right(value)
def failure[A](error: AddonRunConfigError): AddonRunConfigResult[A] = error.toLeft[A]
}
type AddonValidationResult[A] = Either[AddonValidationError, A]
object AddonValidationResult {
def success[A](value: A): AddonValidationResult[A] = Right(value)
def failure[A](error: AddonValidationError): AddonValidationResult[A] = Left(error)
}
def apply[F[_]: Async](
cfg: Config.Addons,
store: Store[F],
userTasks: UserTaskStore[F],
jobStore: JobStore[F],
joex: OJoex[F]
): OAddons[F] =
new OAddons[F] {
private[this] val logger = docspell.logging.getLogger[F]
private val urlReader = FileUrlReader(store.fileRepo)
private val zip = MimeType.zip.asString
private val addonValidate = new AddonValidate[F](cfg, store, joex)
def getAllAddonRunConfigs(collective: Ident): F[List[AddonRunInfo]] =
for {
all <- store.transact(AddonRunConfigData.findAll(collective))
runConfigIDs = all.map(_.runConfig.id).toSet
archiveIds = all.flatMap(_.addons.map(_.addonId)).distinct
archives <- NonEmptyList
.fromList(archiveIds)
.fold(List.empty[RAddonArchive].pure[F])(ids =>
store.transact(RAddonArchive.findByIds(collective, ids))
)
archivesMap = archives.groupBy(_.id)
ptask <- userTasks
.getAll(UserTaskScope.collective(collective))
.filter(ut => runConfigIDs.contains(ut.id))
.map(ut => ut.id -> ut)
.compile
.toList
.map(_.toMap)
result = all.map { t =>
AddonRunInfo.fromRunConfigData(
ptask.get(t.runConfig.id).map(_.timer),
t.addons.map(raa => (archivesMap(raa.addonId).head, raa))
)(t)
}
} yield result
def upsertAddonRunConfig(
collective: Ident,
runConfig: AddonRunInsert
): F[AddonRunConfigResult[Ident]] = {
val insertDataRaw = AddonRunConfigData(
RAddonRunConfig(
runConfig.id,
collective,
runConfig.userId,
runConfig.name,
runConfig.enabled,
Timestamp.Epoch
),
runConfig.addons.zipWithIndex.map { case (a, index) =>
RAddonRunConfigAddon(Ident.unsafe(""), runConfig.id, a.addonId, a.args, index)
}.toList,
runConfig.triggered
.map(t => RAddonRunConfigTrigger(Ident.unsafe(""), runConfig.id, t))
.toList
)
val upsert = for {
userId <-
OptionT
.fromOption(runConfig.userId)
.flatMapF(uid => store.transact(RUser.getIdByIdOrLogin(uid)))
.map(_.uid)
.value
insertData =
insertDataRaw.copy(runConfig =
insertDataRaw.runConfig.copy(userId = userId.orElse(runConfig.userId))
)
id <-
OptionT(store.transact(RAddonRunConfig.findById(collective, runConfig.id)))
.map(rt =>
AddonRunConfigData(
rt.copy(
userId = insertData.runConfig.userId,
name = insertData.runConfig.name,
enabled = insertData.runConfig.enabled
),
insertData.addons,
insertData.triggers
)
)
.semiflatMap(rt =>
store.transact(AddonRunConfigData.update(rt).as(rt.runConfig.id))
)
.getOrElseF(store.transact(AddonRunConfigData.insert(insertData)))
} yield id
EitherT(AddonRunConfigValidate(store, collective)(runConfig))
.semiflatMap(_ =>
upsert.flatTap { runConfigId =>
runConfig.schedule match {
case Some(timer) =>
userTasks.updateTask(
UserTaskScope.collective(collective),
s"Addon task ${runConfig.name}".some,
UserTask(
runConfigId,
scheduledAddonTaskName,
true,
timer,
s"Running scheduled addon task ${runConfig.name}".some,
ScheduledAddonTaskArgs(collective, runConfigId)
)
)
case None =>
userTasks.deleteTask(UserTaskScope.collective(collective), runConfigId)
}
}
)
.value
}
def deleteAddonRunConfig(collective: Ident, runConfigId: Ident): F[Boolean] = {
val deleteRunConfig =
(for {
e <- OptionT(RAddonRunConfig.findById(collective, runConfigId))
_ <- OptionT.liftF(RAddonRunConfigAddon.deleteAllForConfig(e.id))
_ <- OptionT.liftF(RAddonRunConfigTrigger.deleteAllForConfig(e.id))
_ <- OptionT.liftF(RAddonRunConfig.deleteById(collective, e.id))
} yield true).getOrElse(false)
for {
deleted <- store.transact(deleteRunConfig)
_ <-
if (deleted)
userTasks.deleteTask(UserTaskScope.collective(collective), runConfigId)
else 0.pure[F]
} yield deleted
}
def getAllAddons(collective: Ident): F[List[RAddonArchive]] =
store.transact(RAddonArchive.listAll(collective))
def deleteAddon(collective: Ident, addonId: Ident): F[Boolean] =
store.transact(RAddonArchive.deleteById(collective, addonId)).map(_ > 0)
def inspectAddon(
collective: Ident,
url: LenientUri
): F[AddonValidationResult[AddonMeta]] =
addonValidate.fromUrl(collective, url, urlReader, checkExisting = false)
def registerAddon(
collective: Ident,
url: LenientUri,
logger: Option[Logger[F]]
): F[AddonValidationResult[(RAddonArchive, AddonMeta)]] = {
val log = logger.getOrElse(this.logger)
def validateAndInsert(file: FileKey, localUrl: LenientUri) =
addonValidate.fromUrl(collective, url, urlReader, localUrl.some).flatMap {
case Right(meta) =>
insertAddon(collective, url, meta, file)
.map(ar => AddonValidationResult.success(ar -> meta))
case Left(error) =>
store.fileRepo
.delete(file)
.as(AddonValidationResult.failure[(RAddonArchive, AddonMeta)](error))
}
log.info(s"Store addon file from '${url.asString} for ${collective.id}") *>
storeAddonFromUrl(collective, url).flatMapF { file =>
val localUrl = FileUrlReader.url(file)
for {
_ <- log.info(s"Validating addon…")
res <- validateAndInsert(file, localUrl)
_ <- log.info(s"Validation result: $res")
} yield res
}.value
}
def refreshAddon(
collective: Ident,
addonId: Ident
): F[AddonValidationResult[(RAddonArchive, AddonMeta)]] = {
val findAddon = store
.transact(RAddonArchive.findById(collective, addonId))
.map(_.toRight(AddonNotFound))
def validateAddon(aa: RAddonArchive): F[AddonValidationResult[AddonMeta]] =
aa.originalUrl.fold(
AddonValidationResult.failure[AddonMeta](RefreshLocalAddon).pure[F]
)(url =>
addonValidate.fromUrl(collective, url, urlReader, checkExisting = false)
)
EitherT(findAddon).flatMap { aa =>
EitherT(validateAddon(aa))
.flatMap(meta => refreshAddon(aa, meta).map(na => na -> meta))
}.value
}
private def refreshAddon(
r: RAddonArchive,
meta: AddonMeta
): EitherT[F, AddonValidationError, RAddonArchive] =
if (r.isUnchanged(meta)) EitherT.pure(r)
else
r.originalUrl match {
case Some(url) =>
EitherT(
store
.transact(
RAddonArchive
.findByNameAndVersion(r.cid, meta.meta.name, meta.meta.version)
)
.map(
_.fold(().asRight[AddonValidationError])(rx => AddonExists(rx).toLeft)
)
).flatMap(_ =>
storeAddonFromUrl(r.cid, url).flatMap { file =>
val nr = r.update(file, meta)
for {
_ <- EitherT(
store
.transact(RAddonArchive.update(nr))
.map(_.asRight[AddonValidationError])
.recoverWith { case ex =>
logger.warn(ex)(s"Storing addon metadata failed.") *>
store.fileRepo
.delete(file)
.as(
AddonExists(
s"The addon '${nr.name}/${nr.version}' could not be stored",
nr
).toLeft
)
}
)
_ <- EitherT.liftF(store.fileRepo.delete(r.fileId))
} yield nr
}
)
case None =>
EitherT.leftT(RefreshLocalAddon.cast)
}
private def insertAddon(
collective: Ident,
url: LenientUri,
meta: AddonMeta,
file: FileKey
): F[RAddonArchive] =
for {
now <- Timestamp.current[F]
aId <- Ident.randomId[F]
record = RAddonArchive(
aId,
collective,
file,
url.some,
meta,
now
)
_ <- store
.transact(RAddonArchive.insert(record, silent = false))
.onError(_ => store.fileRepo.delete(file))
} yield record
private def storeAddonFromUrl(collective: Ident, url: LenientUri) =
for {
urlFile <- EitherT.pure(url.path.segments.lastOption)
file <- EitherT(
urlReader(url)
.through(
store.fileRepo.save(
collective,
FileCategory.Addon,
MimeTypeHint(urlFile, zip.some)
)
)
.compile
.lastOrError
.attempt
.map(_.leftMap(DownloadFailed(_).cast))
)
} yield file
def runAddonForItem(
account: AccountId,
itemIds: NonEmptyList[Ident],
addonRunConfigIds: Set[Ident]
): F[Unit] =
for {
jobs <- itemIds.traverse(id =>
JobFactory.existingItemAddon(
ItemAddonTaskArgs(account.collective, id, addonRunConfigIds),
account
)
)
_ <- jobStore.insertAllIfNew(jobs.map(_.encode).toList)
} yield ()
}
}

View File

@ -0,0 +1,223 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.backend.ops
import cats.data.{NonEmptyList => Nel, OptionT}
import cats.effect._
import cats.syntax.all._
import fs2.Stream
import docspell.backend.JobFactory
import docspell.common.MakePreviewArgs.StoreMode
import docspell.common._
import docspell.files.TikaMimetype
import docspell.ftsclient.{FtsClient, TextData}
import docspell.scheduler.JobStore
import docspell.store.Store
import docspell.store.queries.QAttachment
import docspell.store.records._
trait OAttachment[F[_]] {
def setExtractedText(
collective: Ident,
itemId: Ident,
attachId: Ident,
newText: F[String]
): F[Unit]
def addOrReplacePdf(
collective: Ident,
attachId: Ident,
pdfData: Stream[F, Byte],
regeneratePreview: Boolean
): F[Unit]
def addOrReplacePreview(
collective: Ident,
attachId: Ident,
imageData: Stream[F, Byte]
): F[Unit]
}
object OAttachment {
def apply[F[_]: Sync](
store: Store[F],
fts: FtsClient[F],
jobStore: JobStore[F]
): OAttachment[F] =
new OAttachment[F] {
private[this] val logger = docspell.logging.getLogger[F]
def setExtractedText(
collective: Ident,
itemId: Ident,
attachId: Ident,
newText: F[String]
): F[Unit] =
for {
_ <- logger.info(s"Find attachment ${attachId.id} to update extracted text.")
cca <- store
.transact(
QAttachment
.allAttachmentMetaAndName(
collective.some,
Nel.of(itemId).some,
ItemState.validStates.append(ItemState.Processing),
100
)
)
.filter(_.id == attachId)
.compile
.last
content = cca.find(_.id == attachId)
_ <- logger.debug(s"Found existing metadata: ${content.isDefined}")
_ <- OptionT
.fromOption(content)
.semiflatMap { cnt =>
for {
_ <- logger.debug(s"Setting new extracted text on ${cnt.id.id}")
text <- newText
td = TextData.attachment(
cnt.item,
cnt.id,
cnt.collective,
cnt.folder,
cnt.lang,
cnt.name,
text.some
)
_ <- store.transact(RAttachmentMeta.updateContent(attachId, text))
_ <- fts.updateIndex(logger, td)
} yield ()
}
.getOrElseF(
logger.warn(
s"Item or attachment meta not found to update text: ${itemId.id}"
)
)
} yield ()
def addOrReplacePdf(
collective: Ident,
attachId: Ident,
pdfData: Stream[F, Byte],
regeneratePreview: Boolean
): F[Unit] = {
def generatePreview(ra: RAttachment): F[Unit] =
JobFactory
.makePreview(MakePreviewArgs(ra.id, StoreMode.Replace), None)
.map(_.encode)
.flatMap(jobStore.insert) *>
logger.info(s"Job submitted to re-generate preview from new pdf")
def generatePageCount(ra: RAttachment): F[Unit] =
JobFactory
.makePageCount(
MakePageCountArgs(ra.id),
AccountId(collective, DocspellSystem.user).some
)
.map(_.encode)
.flatMap(jobStore.insert) *>
logger.info(s"Job submitted to find page count from new pdf")
def setFile(ra: RAttachment, rs: RAttachmentSource) =
for {
_ <- requireMimeType(pdfData, MimeType.pdf)
newFile <- pdfData
.through(
store.fileRepo.save(
collective,
FileCategory.AttachmentConvert,
MimeTypeHint.advertised(MimeType.pdf)
)
)
.compile
.lastOrError
_ <- store.transact(RAttachment.updateFileId(attachId, newFile))
_ <- logger.info(s"Deleting old file for attachment")
_ <-
if (rs.fileId == ra.fileId) ().pure[F]
else store.fileRepo.delete(ra.fileId)
_ <-
if (regeneratePreview) generatePreview(ra)
else ().pure[F]
_ <- generatePageCount(ra)
} yield ()
(for {
ra <- OptionT(
store.transact(RAttachment.findByIdAndCollective(attachId, collective))
)
rs <- OptionT(
store.transact(RAttachmentSource.findByIdAndCollective(attachId, collective))
)
_ <- OptionT.liftF(setFile(ra, rs))
} yield ()).getOrElseF(
logger.warn(
s"Cannot replace pdf file. Attachment not found for id: ${attachId.id}"
)
)
}
def addOrReplacePreview(
collective: Ident,
attachId: Ident,
imageData: Stream[F, Byte]
): F[Unit] = {
def setFile(ra: RAttachment): F[Unit] =
for {
_ <- requireMimeType(imageData, MimeType.image("*"))
newFile <- imageData
.through(
store.fileRepo
.save(collective, FileCategory.PreviewImage, MimeTypeHint.none)
)
.compile
.lastOrError
now <- Timestamp.current[F]
record = RAttachmentPreview(ra.id, newFile, None, now)
oldFile <- store.transact(RAttachmentPreview.upsert(record))
_ <- OptionT
.fromOption(oldFile)
.semiflatMap(store.fileRepo.delete)
.getOrElse(())
} yield ()
(for {
ra <- OptionT(
store.transact(RAttachment.findByIdAndCollective(attachId, collective))
)
_ <- OptionT.liftF(setFile(ra))
} yield ()).getOrElseF(
logger.warn(
s"Cannot add/replace preview file. Attachment not found for id: ${attachId.id}"
)
)
}
}
private def requireMimeType[F[_]: Sync](
data: Stream[F, Byte],
expectedMime: MimeType
): F[Unit] =
TikaMimetype
.detect(data, MimeTypeHint.advertised(expectedMime))
.flatMap { mime =>
if (expectedMime.matches(mime)) ().pure[F]
else
Sync[F].raiseError(
new IllegalArgumentException(
s"Expected pdf file, but got: ${mime.asString}"
)
)
}
}

View File

@ -61,6 +61,12 @@ trait OItem[F[_]] {
collective: Ident
): F[AttachedEvent[UpdateResult]]
def removeTagsOfCategories(
item: Ident,
collective: Ident,
categories: Set[String]
): F[AttachedEvent[UpdateResult]]
def removeTagsMultipleItems(
items: Nel[Ident],
tags: List[String],
@ -80,11 +86,13 @@ trait OItem[F[_]] {
collective: Ident
): F[UpdateResult]
def setFolder(item: Ident, folder: Option[Ident], collective: Ident): F[UpdateResult]
/** Set or remove the folder on an item. Folder can be the id or name. */
def setFolder(item: Ident, folder: Option[String], collective: Ident): F[UpdateResult]
/** Set or remove the folder on multiple items. Folder can be the id or name. */
def setFolderMultiple(
items: Nel[Ident],
folder: Option[Ident],
folder: Option[String],
collective: Ident
): F[UpdateResult]
@ -122,6 +130,13 @@ trait OItem[F[_]] {
def setNotes(item: Ident, notes: Option[String], collective: Ident): F[UpdateResult]
def addNotes(
item: Ident,
notes: String,
separator: Option[String],
collective: Ident
): F[UpdateResult]
def setName(item: Ident, name: String, collective: Ident): F[UpdateResult]
def setNameMultiple(
@ -288,6 +303,28 @@ object OItem {
}
}
def removeTagsOfCategories(
item: Ident,
collective: Ident,
categories: Set[String]
): F[AttachedEvent[UpdateResult]] =
if (categories.isEmpty) {
AttachedEvent.only(UpdateResult.success).pure[F]
} else {
val dbtask =
for {
tags <- RTag.findByItem(item)
removeTags = tags.filter(_.category.exists(categories.contains))
_ <- RTagItem.removeAllTags(item, removeTags.map(_.tagId))
mkEvent = Event.TagsChanged
.partial(Nel.of(item), Nil, removeTags.map(_.tagId.id).toList)
} yield AttachedEvent(UpdateResult.success)(mkEvent)
OptionT(store.transact(RItem.checkByIdAndCollective(item, collective)))
.semiflatMap(_ => store.transact(dbtask))
.getOrElse(AttachedEvent.only(UpdateResult.notFound))
}
def removeTagsMultipleItems(
items: Nel[Ident],
tags: List[String],
@ -420,21 +457,27 @@ object OItem {
def setFolder(
item: Ident,
folder: Option[Ident],
folder: Option[String],
collective: Ident
): F[UpdateResult] =
UpdateResult
.fromUpdate(
store
.transact(RItem.updateFolder(item, collective, folder))
for {
result <- store.transact(RItem.updateFolder(item, collective, folder)).attempt
ures = result.fold(
UpdateResult.failure,
t => UpdateResult.fromUpdateRows(t._1)
)
.flatTap(
onSuccessIgnoreError(fts.updateFolder(logger, item, collective, folder))
_ <- result.fold(
_ => ().pure[F],
t =>
onSuccessIgnoreError(fts.updateFolder(logger, item, collective, t._2))(
ures
)
)
} yield ures
def setFolderMultiple(
items: Nel[Ident],
folder: Option[Ident],
folder: Option[String],
collective: Ident
): F[UpdateResult] =
for {
@ -615,6 +658,33 @@ object OItem {
}
)
def addNotes(
item: Ident,
notes: String,
separator: Option[String],
collective: Ident
): F[UpdateResult] =
store
.transact(RItem.appendNotes(item, collective, notes, separator))
.flatMap {
case Some(newNotes) =>
store
.transact(RCollective.findLanguage(collective))
.map(_.getOrElse(Language.English))
.flatMap(lang =>
fts.updateItemNotes(logger, item, collective, lang, newNotes.some)
)
.attempt
.flatMap {
case Right(()) => ().pure[F]
case Left(ex) =>
logger.warn(s"Error updating full-text index: ${ex.getMessage}")
}
.as(UpdateResult.success)
case None =>
UpdateResult.notFound.pure[F]
}
def setName(item: Ident, name: String, collective: Ident): F[UpdateResult] =
UpdateResult
.fromUpdate(

View File

@ -6,11 +6,13 @@
package docspell.backend.ops
import cats.Applicative
import cats.effect._
import cats.implicits._
import cats.syntax.all._
import fs2.Stream
import docspell.common.Ident
import docspell.common.{Ident, NodeType}
import docspell.joexapi.client.JoexClient
import docspell.joexapi.model.AddonSupport
import docspell.pubsub.api.PubSubT
import docspell.scheduler.msg.{CancelJob, JobsNotify, PeriodicTaskNotify}
@ -21,10 +23,16 @@ trait OJoex[F[_]] {
def notifyPeriodicTasks: F[Unit]
def cancelJob(job: Ident, worker: Ident): F[Unit]
def getAddonSupport: F[List[AddonSupport]]
}
object OJoex {
def apply[F[_]: Applicative](pubSub: PubSubT[F]): Resource[F, OJoex[F]] =
def apply[F[_]: Async](
pubSub: PubSubT[F],
nodes: ONode[F],
joexClient: JoexClient[F]
): Resource[F, OJoex[F]] =
Resource.pure[F, OJoex[F]](new OJoex[F] {
def notifyAllNodes: F[Unit] =
@ -35,5 +43,17 @@ object OJoex {
def cancelJob(job: Ident, worker: Ident): F[Unit] =
pubSub.publish1IgnoreErrors(CancelJob.topic, CancelJob(job, worker)).as(())
def getAddonSupport: F[List[AddonSupport]] =
for {
joex <- nodes.getNodes(NodeType.Joex)
conc = math.max(2, Runtime.getRuntime.availableProcessors() - 1)
supp <- Stream
.emits(joex)
.covary[F]
.parEvalMap(conc)(n => joexClient.getAddonSupport(n.url))
.compile
.toList
} yield supp
})
}

View File

@ -13,11 +13,27 @@ import docspell.common.{Ident, LenientUri, NodeType}
import docspell.store.Store
import docspell.store.records.RNode
import scodec.bits.ByteVector
trait ONode[F[_]] {
def register(appId: Ident, nodeType: NodeType, uri: LenientUri): F[Unit]
def register(
appId: Ident,
nodeType: NodeType,
uri: LenientUri,
serverSecret: Option[ByteVector]
): F[Unit]
def unregister(appId: Ident): F[Unit]
def withRegistered(
appId: Ident,
nodeType: NodeType,
uri: LenientUri,
serverSecret: Option[ByteVector]
): Resource[F, Unit]
def getNodes(nodeType: NodeType): F[Vector[RNode]]
}
object ONode {
@ -25,9 +41,14 @@ object ONode {
def apply[F[_]: Async](store: Store[F]): Resource[F, ONode[F]] =
Resource.pure[F, ONode[F]](new ONode[F] {
val logger = docspell.logging.getLogger[F]
def register(appId: Ident, nodeType: NodeType, uri: LenientUri): F[Unit] =
def register(
appId: Ident,
nodeType: NodeType,
uri: LenientUri,
serverSecret: Option[ByteVector]
): F[Unit] =
for {
node <- RNode(appId, nodeType, uri)
node <- RNode(appId, nodeType, uri, serverSecret)
_ <- logger.info(s"Registering node ${node.id.id}")
_ <- store.transact(RNode.set(node))
} yield ()
@ -35,6 +56,19 @@ object ONode {
def unregister(appId: Ident): F[Unit] =
logger.info(s"Unregister app ${appId.id}") *>
store.transact(RNode.delete(appId)).map(_ => ())
def withRegistered(
appId: Ident,
nodeType: NodeType,
uri: LenientUri,
serverSecret: Option[ByteVector]
): Resource[F, Unit] =
Resource.make(register(appId, nodeType, uri, serverSecret))(_ =>
unregister(appId)
)
def getNodes(nodeType: NodeType): F[Vector[RNode]] =
store.transact(RNode.findAll(nodeType))
})
}

View File

@ -9,8 +9,9 @@ package docspell.common
import java.time.Instant
import io.circe._
import scodec.bits.ByteVector
object BaseJsonCodecs {
trait BaseJsonCodecs {
implicit val encodeInstantEpoch: Encoder[Instant] =
Encoder.encodeJavaLong.contramap(_.toEpochMilli)
@ -18,4 +19,11 @@ object BaseJsonCodecs {
implicit val decodeInstantEpoch: Decoder[Instant] =
Decoder.decodeLong.map(Instant.ofEpochMilli)
implicit val byteVectorEncoder: Encoder[ByteVector] =
Encoder.encodeString.contramap(_.toBase64)
implicit val byteVectorDecoder: Decoder[ByteVector] =
Decoder.decodeString.emap(ByteVector.fromBase64Descriptive(_))
}
object BaseJsonCodecs extends BaseJsonCodecs

View File

@ -18,6 +18,18 @@ final case class Binary[F[_]](name: String, mime: MimeType, data: Stream[F, Byte
def withMime(mime: MimeType): Binary[F] =
copy(mime = mime)
/** Return the extension of `name` if available (without the dot) */
def extension: Option[String] =
name.lastIndexOf('.') match {
case n if n > 0 =>
Some(name.substring(n + 1))
case _ =>
None
}
def extensionIn(extensions: Set[String]): Boolean =
extension.exists(extensions.contains)
}
object Binary {

View File

@ -32,6 +32,7 @@ object FileCategory {
case object PreviewImage extends FileCategory
case object Classifier extends FileCategory
case object DownloadAll extends FileCategory
case object Addon extends FileCategory
val all: NonEmptyList[FileCategory] =
NonEmptyList.of(
@ -39,7 +40,8 @@ object FileCategory {
AttachmentConvert,
PreviewImage,
Classifier,
DownloadAll
DownloadAll,
Addon
)
def fromString(str: String): Either[String, FileCategory] =

View File

@ -32,7 +32,8 @@ object Glob {
def single(str: String) =
PatternGlob(Pattern(split(str, separator).map(makeSegment)))
if (in == "*") all
if (in == all.asString) all
else if (in == none.asString) none
else
split(in, anyChar) match {
case NonEmptyList(_, Nil) =>
@ -51,15 +52,25 @@ object Glob {
val asString = "*"
}
val none = new Glob {
def matches(caseSensitive: Boolean)(in: String) = false
def matchFilenameOrPath(in: String) = false
def asString = "!*"
}
def pattern(pattern: Pattern): Glob =
PatternGlob(pattern)
/** A simple glob supporting `*` and `?`. */
final private case class PatternGlob(pattern: Pattern) extends Glob {
def matches(caseSensitive: Boolean)(in: String): Boolean =
def matches(caseSensitive: Boolean)(in: String): Boolean = {
val input = Glob.split(in, Glob.separator)
pattern.parts.size == input.size &&
pattern.parts
.zipWith(Glob.split(in, Glob.separator))(_.matches(caseSensitive)(_))
.zipWith(input)(_.matches(caseSensitive)(_))
.forall(identity)
}
def matchFilenameOrPath(in: String): Boolean =
if (pattern.parts.tail.isEmpty) matches(true)(split(in, separator).last)
@ -67,6 +78,8 @@ object Glob {
def asString: String =
pattern.asString
override def toString = s"PatternGlob($asString)"
}
final private case class AnyGlob(globs: NonEmptyList[Glob]) extends Glob {
@ -76,6 +89,8 @@ object Glob {
globs.exists(_.matchFilenameOrPath(in))
def asString =
globs.toList.map(_.asString).mkString(anyChar.toString)
override def toString = s"AnyGlob($globs)"
}
case class Pattern(parts: NonEmptyList[Segment]) {

View File

@ -26,6 +26,9 @@ case class Ident(id: String) {
def /(next: Ident): Ident =
new Ident(id + Ident.concatChar + next.id)
def take(n: Int): Ident =
new Ident(id.take(n))
}
object Ident {

View File

@ -0,0 +1,28 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.common
import io.circe.generic.semiauto.{deriveDecoder, deriveEncoder}
import io.circe.{Decoder, Encoder}
/** Arguments to submit a task that runs addons configured for some existing item.
*
* If `addonTaskIds` is non empty, only these addon tasks are run. Otherwise all addon
* tasks that are configured for 'existing-item' are run.
*/
final case class ItemAddonTaskArgs(
collective: Ident,
itemId: Ident,
addonRunConfigs: Set[Ident]
)
object ItemAddonTaskArgs {
val taskName: Ident = Ident.unsafe("addon-existing-item")
implicit val jsonDecoder: Decoder[ItemAddonTaskArgs] = deriveDecoder
implicit val jsonEncoder: Encoder[ItemAddonTaskArgs] = deriveEncoder
}

View File

@ -6,6 +6,8 @@
package docspell.common
import fs2.io.file.Path
case class MimeTypeHint(filename: Option[String], advertised: Option[String]) {
def withName(name: String): MimeTypeHint =
@ -21,6 +23,9 @@ object MimeTypeHint {
def filename(name: String): MimeTypeHint =
MimeTypeHint(Some(name), None)
def filename(file: Path): MimeTypeHint =
filename(file.fileName.toString)
def advertised(mimeType: MimeType): MimeTypeHint =
advertised(mimeType.asString)

View File

@ -17,7 +17,7 @@ import io.circe.generic.semiauto._
* This task is run for each new file to create a new item from it or to add this file as
* an attachment to an existing item.
*
* If the `itemId' is set to some value, the item is tried to load to ammend with the
* If the `itemId' is set to some value, the item is tried to load to amend with the
* given files. Otherwise a new item is created.
*
* It is also re-used by the 'ReProcessItem' task.

View File

@ -0,0 +1,19 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.common
import io.circe.generic.semiauto.{deriveDecoder, deriveEncoder}
import io.circe.{Decoder, Encoder}
final case class ScheduledAddonTaskArgs(collective: Ident, addonTaskId: Ident)
object ScheduledAddonTaskArgs {
val taskName: Ident = Ident.unsafe("addon-scheduled-task")
implicit val jsonDecoder: Decoder[ScheduledAddonTaskArgs] = deriveDecoder
implicit val jsonEncoder: Encoder[ScheduledAddonTaskArgs] = deriveEncoder
}

View File

@ -17,11 +17,23 @@ import cats.implicits._
import fs2.io.file.Path
import fs2.{Stream, io, text}
import docspell.common.{exec => newExec}
import docspell.logging.Logger
// better use `SysCmd` and `SysExec`
object SystemCommand {
final case class Config(program: String, args: Seq[String], timeout: Duration) {
final case class Config(
program: String,
args: Seq[String],
timeout: Duration,
env: Map[String, String] = Map.empty
) {
def toSysCmd = newExec
.SysCmd(program, newExec.Args(args))
.withTimeout(timeout)
.addEnv(newExec.Env(env))
def mapArgs(f: String => String): Config =
Config(program, args.map(f), timeout)
@ -33,6 +45,18 @@ object SystemCommand {
}
)
def withEnv(key: String, value: String): Config =
copy(env = env.updated(key, value))
def addEnv(moreEnv: Map[String, String]): Config =
copy(env = env ++ moreEnv)
def appendArgs(extraArgs: Args): Config =
copy(args = args ++ extraArgs.args)
def appendArgs(extraArgs: Seq[String]): Config =
copy(args = args ++ extraArgs)
def toCmd: List[String] =
program :: args.toList
@ -40,6 +64,45 @@ object SystemCommand {
toCmd.mkString(" ")
}
final case class Args(args: Vector[String]) extends Iterable[String] {
override def iterator = args.iterator
def prepend(a: String): Args = Args(a +: args)
def prependWhen(flag: Boolean)(a: String): Args =
prependOption(Option.when(flag)(a))
def prependOption(value: Option[String]): Args =
value.map(prepend).getOrElse(this)
def append(a: String, as: String*): Args =
Args(args ++ (a +: as.toVector))
def appendOption(value: Option[String]): Args =
value.map(append(_)).getOrElse(this)
def appendOptionVal(first: String, second: Option[String]): Args =
second.map(b => append(first, b)).getOrElse(this)
def appendWhen(flag: Boolean)(a: String, as: String*): Args =
if (flag) append(a, as: _*) else this
def appendWhenNot(flag: Boolean)(a: String, as: String*): Args =
if (!flag) append(a, as: _*) else this
def append(p: Path): Args =
append(p.toString)
def append(as: Iterable[String]): Args =
Args(args ++ as.toVector)
}
object Args {
val empty: Args = Args()
def apply(as: String*): Args =
Args(as.toVector)
}
final case class Result(rc: Int, stdout: String, stderr: String)
def exec[F[_]: Sync](
@ -104,6 +167,10 @@ object SystemCommand {
.redirectError(Redirect.PIPE)
.redirectOutput(Redirect.PIPE)
val pbEnv = pb.environment()
cmd.env.foreach { case (key, value) =>
pbEnv.put(key, value)
}
wd.map(_.toNioPath.toFile).foreach(pb.directory)
pb.start()
}

View File

@ -0,0 +1,94 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.common
import cats.data.NonEmptyList
import cats.kernel.Monoid
import cats.syntax.all._
trait UrlMatcher {
def matches(url: LenientUri): Boolean
}
object UrlMatcher {
val True = instance(_ => true)
val False = instance(_ => false)
def instance(f: LenientUri => Boolean): UrlMatcher =
(url: LenientUri) => f(url)
def fromString(str: String): Either[String, UrlMatcher] =
if (str == "") False.asRight
else if (str == "*") True.asRight
else LenientUri.parse(str).map(fromUrl)
def unsafeFromString(str: String): UrlMatcher =
fromString(str).fold(sys.error, identity)
def fromStringList(str: List[String]): Either[String, UrlMatcher] =
str match {
case Nil => False.asRight
case _ => str.map(_.trim).traverse(fromString).map(_.combineAll)
}
def fromUrl(url: LenientUri): UrlMatcher = {
val schemeGlob = Glob(url.scheme.head)
val hostGlob = HostGlob(url.host)
val pathGlob = Glob(url.path.asString)
new Impl(schemeGlob, hostGlob, pathGlob, url.path.segments.size)
}
def any(ulrm: IterableOnce[UrlMatcher]): UrlMatcher =
anyMonoid.combineAll(ulrm)
def all(urlm: IterableOnce[UrlMatcher]): UrlMatcher =
allMonoid.combineAll(urlm)
val anyMonoid: Monoid[UrlMatcher] =
Monoid.instance(False, (a, b) => instance(url => a.matches(url) || b.matches(url)))
val allMonoid: Monoid[UrlMatcher] =
Monoid.instance(True, (a, b) => instance(url => a.matches(url) && b.matches(url)))
implicit val defaultMonoid: Monoid[UrlMatcher] = anyMonoid
private class Impl(scheme: Glob, host: HostGlob, path: Glob, pathSegmentCount: Int)
extends UrlMatcher {
def matches(url: LenientUri) = {
// strip path to only match prefixes
val mPath: LenientUri.Path =
NonEmptyList.fromList(url.path.segments.take(pathSegmentCount)) match {
case Some(nel) => LenientUri.NonEmptyPath(nel)
case None => LenientUri.RootPath
}
url.scheme.forall(scheme.matches(false)) &&
host.matches(url.host) &&
path.matchFilenameOrPath(mPath.asString)
}
}
private class HostGlob(glob: Option[Glob]) {
def matches(host: Option[String]): Boolean =
(glob, host) match {
case (Some(pattern), Some(word)) =>
pattern.matches(false)(HostGlob.prepareHost(word))
case (None, None) => true
case _ => false
}
override def toString = s"HostGlob(${glob.map(_.asString)})"
}
private object HostGlob {
def apply(hostPattern: Option[String]): HostGlob =
new HostGlob(hostPattern.map(p => Glob(prepareHost(p))))
private def prepareHost(host: String): String =
host.replace('.', '/')
}
}

View File

@ -0,0 +1,35 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.common
import cats.ApplicativeError
import cats.effect._
import fs2.Stream
trait UrlReader[F[_]] {
def apply(url: LenientUri): Stream[F, Byte]
}
object UrlReader {
def instance[F[_]](f: LenientUri => Stream[F, Byte]): UrlReader[F] =
(url: LenientUri) => f(url)
def failWith[F[_]](
message: String
)(implicit F: ApplicativeError[F, Throwable]): UrlReader[F] =
instance(url =>
Stream.raiseError(
new IllegalStateException(s"Unable to read '${url.asString}': $message")
)
)
def apply[F[_]](implicit r: UrlReader[F]): UrlReader[F] = r
implicit def defaultReader[F[_]: Sync]: UrlReader[F] =
instance(_.readURL[F](8192))
}

View File

@ -0,0 +1,30 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.common.bc
import io.circe.generic.extras.Configuration
import io.circe.generic.extras.semiauto.{deriveConfiguredDecoder, deriveConfiguredEncoder}
import io.circe.generic.semiauto.{deriveDecoder, deriveEncoder}
import io.circe.{Decoder, Encoder}
sealed trait AttachmentAction {}
object AttachmentAction {
implicit val deriveConfig: Configuration =
Configuration.default.withDiscriminator("action").withKebabCaseConstructorNames
case class SetExtractedText(text: Option[String]) extends AttachmentAction
object SetExtractedText {
implicit val jsonDecoder: Decoder[SetExtractedText] = deriveDecoder
implicit val jsonEncoder: Encoder[SetExtractedText] = deriveEncoder
}
implicit val jsonDecoder: Decoder[AttachmentAction] = deriveConfiguredDecoder
implicit val jsonEncoder: Encoder[AttachmentAction] = deriveConfiguredEncoder
}

View File

@ -0,0 +1,44 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.common.bc
import docspell.common.Ident
import io.circe.generic.extras.Configuration
import io.circe.generic.extras.semiauto.{deriveConfiguredDecoder, deriveConfiguredEncoder}
import io.circe.generic.semiauto.{deriveDecoder, deriveEncoder}
import io.circe.{Decoder, Encoder}
sealed trait BackendCommand {}
object BackendCommand {
implicit val deriveConfig: Configuration =
Configuration.default.withDiscriminator("command").withKebabCaseConstructorNames
case class ItemUpdate(itemId: Ident, actions: List[ItemAction]) extends BackendCommand
object ItemUpdate {
implicit val jsonDecoder: Decoder[ItemUpdate] = deriveDecoder
implicit val jsonEncoder: Encoder[ItemUpdate] = deriveEncoder
}
def item(itemId: Ident, actions: List[ItemAction]): BackendCommand =
ItemUpdate(itemId, actions)
case class AttachmentUpdate(
itemId: Ident,
attachId: Ident,
actions: List[AttachmentAction]
) extends BackendCommand
object AttachmentUpdate {
implicit val jsonDecoder: Decoder[AttachmentUpdate] = deriveDecoder
implicit val jsonEncoder: Encoder[AttachmentUpdate] = deriveEncoder
}
implicit val jsonDecoder: Decoder[BackendCommand] = deriveConfiguredDecoder
implicit val jsonEncoder: Encoder[BackendCommand] = deriveConfiguredEncoder
}

View File

@ -0,0 +1,17 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.common.bc
import docspell.common.Ident
trait BackendCommandRunner[F[_], A] {
def run(collective: Ident, cmd: BackendCommand): F[A]
def runAll(collective: Ident, cmds: List[BackendCommand]): F[A]
}

View File

@ -0,0 +1,102 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.common.bc
import docspell.common.Ident
import io.circe.generic.extras.Configuration
import io.circe.generic.extras.semiauto.{deriveConfiguredDecoder, deriveConfiguredEncoder}
import io.circe.generic.semiauto.{deriveDecoder, deriveEncoder}
import io.circe.{Decoder, Encoder}
sealed trait ItemAction {}
object ItemAction {
implicit val deriveConfig: Configuration =
Configuration.default.withDiscriminator("action").withKebabCaseConstructorNames
case class AddTags(tags: Set[String]) extends ItemAction
object AddTags {
implicit val jsonDecoder: Decoder[AddTags] = deriveDecoder
implicit val jsonEncoder: Encoder[AddTags] = deriveEncoder
}
case class ReplaceTags(tags: Set[String]) extends ItemAction
object ReplaceTags {
implicit val jsonDecoder: Decoder[ReplaceTags] = deriveDecoder
implicit val jsonEncoder: Encoder[ReplaceTags] = deriveEncoder
}
case class RemoveTags(tags: Set[String]) extends ItemAction
object RemoveTags {
implicit val jsonDecoder: Decoder[RemoveTags] = deriveDecoder
implicit val jsonEncoder: Encoder[RemoveTags] = deriveEncoder
}
case class RemoveTagsCategory(categories: Set[String]) extends ItemAction
object RemoveTagsCategory {
implicit val jsonDecoder: Decoder[RemoveTagsCategory] = deriveDecoder
implicit val jsonEncoder: Encoder[RemoveTagsCategory] = deriveEncoder
}
case class SetFolder(folder: Option[String]) extends ItemAction
object SetFolder {
implicit val jsonDecoder: Decoder[SetFolder] = deriveDecoder
implicit val jsonEncoder: Encoder[SetFolder] = deriveEncoder
}
case class SetCorrOrg(id: Option[Ident]) extends ItemAction
object SetCorrOrg {
implicit val jsonDecoder: Decoder[SetCorrOrg] = deriveDecoder
implicit val jsonEncoder: Encoder[SetCorrOrg] = deriveEncoder
}
case class SetCorrPerson(id: Option[Ident]) extends ItemAction
object SetCorrPerson {
implicit val jsonDecoder: Decoder[SetCorrPerson] = deriveDecoder
implicit val jsonEncoder: Encoder[SetCorrPerson] = deriveEncoder
}
case class SetConcPerson(id: Option[Ident]) extends ItemAction
object SetConcPerson {
implicit val jsonDecoder: Decoder[SetConcPerson] = deriveDecoder
implicit val jsonEncoder: Encoder[SetConcPerson] = deriveEncoder
}
case class SetConcEquipment(id: Option[Ident]) extends ItemAction
object SetConcEquipment {
implicit val jsonDecoder: Decoder[SetConcEquipment] = deriveDecoder
implicit val jsonEncoder: Encoder[SetConcEquipment] = deriveEncoder
}
case class SetField(field: Ident, value: String) extends ItemAction
object SetField {
implicit val jsonDecoder: Decoder[SetField] = deriveDecoder
implicit val jsonEncoder: Encoder[SetField] = deriveEncoder
}
case class SetName(name: String) extends ItemAction
object SetName {
implicit val jsonDecoder: Decoder[SetName] = deriveDecoder
implicit val jsonEncoder: Encoder[SetName] = deriveEncoder
}
case class SetNotes(notes: Option[String]) extends ItemAction
object SetNotes {
implicit val jsonDecoder: Decoder[SetNotes] = deriveDecoder
implicit val jsonEncoder: Encoder[SetNotes] = deriveEncoder
}
case class AddNotes(notes: String, separator: Option[String]) extends ItemAction
object AddNotes {
implicit val jsonDecoder: Decoder[AddNotes] = deriveDecoder
implicit val jsonEncoder: Encoder[AddNotes] = deriveEncoder
}
implicit val jsonDecoder: Decoder[ItemAction] = deriveConfiguredDecoder
implicit val jsonEncoder: Encoder[ItemAction] = deriveConfiguredEncoder
}

View File

@ -0,0 +1,49 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.common.exec
import fs2.io.file.Path
case class Args(values: Seq[String]) {
def option(key: String, value: String): Args =
Args(values ++ Seq(key, value))
def option(key: String, value: Option[String]): Args =
value.map(v => option(key, v)).getOrElse(this)
def appendOpt(v: Option[String]): Args =
v.map(e => Args(values :+ e)).getOrElse(this)
def append(v: String, vs: String*): Args =
Args(values ++ (v +: vs))
def append(path: Path): Args =
append(path.toString)
def append(args: Args): Args =
Args(values ++ args.values)
def append(args: Seq[String]): Args =
Args(values ++ args)
def prepend(v: String): Args =
Args(v +: values)
def prependWhen(flag: Boolean)(v: String) =
if (flag) prepend(v) else this
def cmdString: String =
values.mkString(" ")
}
object Args {
val empty: Args = Args(Seq.empty)
def of(v: String*): Args =
Args(v)
}

View File

@ -0,0 +1,37 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.common.exec
case class Env(values: Map[String, String]) {
def add(name: String, value: String): Env =
copy(values.updated(name, value))
def addAll(v: Map[String, String]): Env =
Env(values ++ v)
def addAll(e: Env): Env =
Env(values ++ e.values)
def ++(e: Env) = addAll(e)
def foreach(f: (String, String) => Unit): Unit =
values.foreach(t => f(t._1, t._2))
def map[A](f: (String, String) => A): Seq[A] =
values.map(f.tupled).toSeq
def mapConcat[A](f: (String, String) => Seq[A]): Seq[A] =
values.flatMap(f.tupled).toSeq
}
object Env {
val empty: Env = Env(Map.empty)
def of(nv: (String, String)*): Env =
Env(Map(nv: _*))
}

View File

@ -0,0 +1,43 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.common.exec
import docspell.common._
final case class SysCmd(
program: String,
args: Args,
env: Env,
timeout: Duration
) {
def withArgs(f: Args => Args): SysCmd =
copy(args = f(args))
def withTimeout(to: Duration): SysCmd =
copy(timeout = to)
def withEnv(f: Env => Env): SysCmd =
copy(env = f(env))
def addEnv(env: Env): SysCmd =
withEnv(_.addAll(env))
def cmdString: String =
s"$program ${args.cmdString}"
private[exec] def toCmd: Seq[String] =
program +: args.values
}
object SysCmd {
def apply(prg: String, args: String*): SysCmd =
apply(prg, Args(args))
def apply(prg: String, args: Args): SysCmd =
SysCmd(prg, args, Env.empty, Duration.minutes(2))
}

View File

@ -0,0 +1,163 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.common.exec
import java.lang.ProcessBuilder.Redirect
import java.util.concurrent.TimeUnit
import scala.concurrent.TimeoutException
import scala.jdk.CollectionConverters._
import cats.effect._
import cats.syntax.all._
import fs2.io.file.Path
import fs2.{Pipe, Stream}
import docspell.common.Duration
import docspell.logging.Logger
trait SysExec[F[_]] {
def stdout: Stream[F, Byte]
def stdoutLines: Stream[F, String] =
stdout
.through(fs2.text.utf8.decode)
.through(fs2.text.lines)
def stderr: Stream[F, Byte]
def stderrLines: Stream[F, String] =
stderr
.through(fs2.text.utf8.decode)
.through(fs2.text.lines)
def waitFor(timeout: Option[Duration] = None): F[Int]
/** Sends a signal to the process to terminate it immediately */
def cancel: F[Unit]
/** Consume lines of output of the process in background. */
def consumeOutputs(out: Pipe[F, String, Unit], err: Pipe[F, String, Unit])(implicit
F: Async[F]
): Resource[F, SysExec[F]]
/** Consumes stderr lines (left) and stdout lines (right) in a background thread. */
def consumeOutputs(
m: Either[String, String] => F[Unit]
)(implicit F: Async[F]): Resource[F, SysExec[F]] = {
val pe: Pipe[F, String, Unit] = _.map(_.asLeft).evalMap(m)
val po: Pipe[F, String, Unit] = _.map(_.asRight).evalMap(m)
consumeOutputs(po, pe)
}
def logOutputs(logger: Logger[F], name: String)(implicit F: Async[F]) =
consumeOutputs {
case Right(line) => logger.debug(s"[$name (out)]: $line")
case Left(line) => logger.debug(s"[$name (err)]: $line")
}
}
object SysExec {
private val readChunkSz = 8 * 1024
def apply[F[_]: Sync](
cmd: SysCmd,
logger: Logger[F],
workdir: Option[Path] = None,
stdin: Option[Stream[F, Byte]] = None
): Resource[F, SysExec[F]] =
for {
proc <- startProcess(logger, cmd, workdir, stdin)
fibers <- Resource.eval(Ref.of[F, List[F[Unit]]](Nil))
} yield new SysExec[F] {
def stdout: Stream[F, Byte] =
fs2.io.readInputStream(
Sync[F].blocking(proc.getInputStream),
readChunkSz,
closeAfterUse = false
)
def stderr: Stream[F, Byte] =
fs2.io.readInputStream(
Sync[F].blocking(proc.getErrorStream),
readChunkSz,
closeAfterUse = false
)
def cancel = Sync[F].blocking(proc.destroy())
def waitFor(timeout: Option[Duration]): F[Int] = {
val to = timeout.getOrElse(cmd.timeout)
logger.trace("Waiting for command to terminate…") *>
Sync[F]
.blocking(proc.waitFor(to.millis, TimeUnit.MILLISECONDS))
.flatTap(_ => fibers.get.flatMap(_.traverse_(identity)))
.flatMap(terminated =>
if (terminated) proc.exitValue().pure[F]
else
Sync[F]
.raiseError(
new TimeoutException(s"Timed out after: ${to.formatExact}")
)
)
}
def consumeOutputs(out: Pipe[F, String, Unit], err: Pipe[F, String, Unit])(implicit
F: Async[F]
): Resource[F, SysExec[F]] =
for {
f1 <- F.background(stdoutLines.through(out).compile.drain)
f2 <- F.background(stderrLines.through(err).compile.drain)
_ <- Resource.eval(fibers.update(list => f1.void :: f2.void :: list))
} yield this
}
private def startProcess[F[_]: Sync, A](
logger: Logger[F],
cmd: SysCmd,
workdir: Option[Path],
stdin: Option[Stream[F, Byte]]
): Resource[F, Process] = {
val log = logger.debug(s"Running external command: ${cmd.cmdString}")
val proc = log *>
Sync[F].blocking {
val pb = new ProcessBuilder(cmd.toCmd.asJava)
.redirectInput(if (stdin.isDefined) Redirect.PIPE else Redirect.INHERIT)
.redirectError(Redirect.PIPE)
.redirectOutput(Redirect.PIPE)
val pbEnv = pb.environment()
cmd.env.foreach { (name, v) =>
pbEnv.put(name, v)
()
}
workdir.map(_.toNioPath.toFile).foreach(pb.directory)
pb.start()
}
Resource
.make(proc)(p =>
logger.debug(s"Closing process: `${cmd.cmdString}`").map(_ => p.destroy())
)
.evalMap(p =>
stdin match {
case Some(in) =>
writeToProcess(in, p).compile.drain.as(p)
case None =>
p.pure[F]
}
)
}
private def writeToProcess[F[_]: Sync](
data: Stream[F, Byte],
proc: Process
): Stream[F, Nothing] =
data.through(fs2.io.writeOutputStream(Sync[F].blocking(proc.getOutputStream)))
}

View File

@ -4,20 +4,18 @@
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.common
package docspell.common.util
import java.nio.file.{Path => JPath}
import cats.FlatMap
import cats.Monad
import cats.effect._
import cats.implicits._
import cats.syntax.all._
import cats.{FlatMap, Monad}
import fs2.Stream
import fs2.io.file.{Files, Flags, Path}
import docspell.common.syntax.all._
import io.circe.Decoder
import io.circe.parser
object File {
@ -75,6 +73,5 @@ object File {
.map(_ => file)
def readJson[F[_]: Async, A](file: Path)(implicit d: Decoder[A]): F[A] =
readText[F](file).map(_.parseJsonAs[A]).rethrow
readText[F](file).map(parser.decode[A]).rethrow
}

View File

@ -0,0 +1,27 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.common.util
import cats.effect._
import scodec.bits.ByteVector
trait Random[F[_]] {
def string(len: Int): F[String]
def string: F[String] = string(8)
}
object Random {
def apply[F[_]: Sync] =
new Random[F] {
def string(len: Int) = Sync[F].delay {
val buf = Array.ofDim[Byte](len)
new scala.util.Random().nextBytes(buf)
ByteVector.view(buf).toBase58
}
}
}

View File

@ -70,11 +70,13 @@ class GlobTest extends FunSuite {
test("with splitting") {
assert(Glob("a/b/*").matches(true)("a/b/hello"))
assert(!Glob("a/b/*").matches(true)("a/b/hello/bello"))
assert(!Glob("a/b/*").matches(true)("/a/b/hello"))
assert(Glob("/a/b/*").matches(true)("/a/b/hello"))
assert(!Glob("/a/b/*").matches(true)("a/b/hello"))
assert(!Glob("*/a/b/*").matches(true)("a/b/hello"))
assert(Glob("*/a/b/*").matches(true)("test/a/b/hello"))
assert(!Glob("/a/b").matches(true)("/a/b/c/d"))
}
test("asString") {

View File

@ -0,0 +1,60 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.common
import munit._
class UrlMatcherTest extends FunSuite {
test("it should match patterns") {
assertUrlsMatch(
uri("https://github.com/docspell/*") -> uri("https://github.com/docspell/dsc"),
uri("*s://test.com/*") -> uri("https://test.com/a"),
uri("*s://test.com/*") -> uri("https://test.com/a/b"),
uri("*s://test.com/*") -> uri("https://test.com/a/b/c"),
uri("*s://test.com/project/*") -> uri("https://test.com/project/c"),
uri("https://*.test.com/projects/*") -> uri("https://a.test.com/projects/p1"),
uri("https://*.test.com/projects/*") -> uri("https://b.test.com/projects/p1"),
uri("https://*.test.com/projects/*") -> uri("https://b.test.com/projects/p1")
)
assertUrlsNotMatch(
uri("https://*.test.com/projects/*") -> uri("https://test.com/projects/p1"),
uri("*s://test.com/project/*") -> uri("https://test.com/subject/c")
)
}
def uri(str: String): LenientUri = LenientUri.unsafe(str)
def assertUrlsMatch(tests: List[(LenientUri, LenientUri)]): Unit =
tests.foreach { case (patternUri, checkUri) =>
assert(
UrlMatcher.fromUrl(patternUri).matches(checkUri),
s"$patternUri does not match $checkUri"
)
}
def assertUrlsMatch(
test: (LenientUri, LenientUri),
more: (LenientUri, LenientUri)*
): Unit =
assertUrlsMatch(test :: more.toList)
def assertUrlsNotMatch(tests: List[(LenientUri, LenientUri)]): Unit =
tests.foreach { case (patternUri, checkUri) =>
assert(
!UrlMatcher.fromUrl(patternUri).matches(checkUri),
s"$patternUri incorrectly matches $checkUri"
)
}
def assertUrlsNotMatch(
test: (LenientUri, LenientUri),
more: (LenientUri, LenientUri)*
): Unit =
assertUrlsNotMatch(test :: more.toList)
}

View File

@ -0,0 +1,85 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.common.bc
import docspell.common._
import io.circe.parser
import io.circe.syntax._
import munit._
class BackendCommandTest extends FunSuite {
test("encode json") {
val bc: BackendCommand =
BackendCommand.item(
id("abc"),
List(
ItemAction.RemoveTagsCategory(Set("doctype")),
ItemAction.AddTags(Set("tag1", "tag2"))
)
)
assertEquals(
bc.asJson.spaces2,
"""{
| "itemId" : "abc",
| "actions" : [
| {
| "categories" : [
| "doctype"
| ],
| "action" : "remove-tags-category"
| },
| {
| "tags" : [
| "tag1",
| "tag2"
| ],
| "action" : "add-tags"
| }
| ],
| "command" : "item-update"
|}""".stripMargin
)
}
test("decode case insensitive keys") {
val json = """{
| "itemId" : "abc",
| "actions" : [
| {
| "categories" : [
| "doctype"
| ],
| "action" : "remove-tags-category"
| },
| {
| "tags" : [
| "tag1",
| "tag2"
| ],
| "action" : "add-tags"
| }
| ],
| "command" : "item-update"
|}""".stripMargin
val bc: BackendCommand =
BackendCommand.item(
id("abc"),
List(
ItemAction.RemoveTagsCategory(Set("doctype")),
ItemAction.AddTags(Set("tag1", "tag2"))
)
)
assertEquals(parser.decode[BackendCommand](json), Right(bc))
}
def id(str: String) = Ident.unsafe(str)
}

View File

@ -13,6 +13,7 @@ import scala.reflect.ClassTag
import cats.syntax.all._
import fs2.io.file.Path
import docspell.addons.RunnerType
import docspell.common._
import docspell.ftspsql.{PgQueryParser, RankNormalization}
import docspell.logging.{Level, LogConfig}
@ -32,6 +33,17 @@ object Implicits {
else super.fieldValue(name)
}
implicit val urlMatcherReader: ConfigReader[UrlMatcher] = {
val fromList = ConfigReader[List[String]].emap(reason(UrlMatcher.fromStringList))
val fromString = ConfigReader[String].emap(
reason(str => UrlMatcher.fromStringList(str.split("[\\s,]+").toList))
)
fromList.orElse(fromString)
}
implicit val runnerSelectReader: ConfigReader[List[RunnerType]] =
ConfigReader[String].emap(reason(RunnerType.fromSeparatedString))
implicit val accountIdReader: ConfigReader[AccountId] =
ConfigReader[String].emap(reason(AccountId.parse))

View File

@ -12,6 +12,7 @@ import fs2.io.file.{Files, Path}
import fs2.{Pipe, Stream}
import docspell.common._
import docspell.common.util.File
import docspell.convert.ConversionResult
import docspell.convert.ConversionResult.{Handler, successPdf, successPdfTxt}
import docspell.logging.Logger

View File

@ -15,6 +15,7 @@ import cats.implicits._
import fs2.Stream
import docspell.common._
import docspell.common.util.File
import docspell.convert.ConversionResult.Handler
import docspell.convert.extern.OcrMyPdfConfig
import docspell.convert.extern.{TesseractConfig, UnoconvConfig, WkHtmlPdfConfig}

View File

@ -18,6 +18,7 @@ import fs2.io.file.Path
import fs2.{Pipe, Stream}
import docspell.common._
import docspell.common.util.File
import docspell.convert.ConversionResult.Handler
import docspell.files.TikaMimetype

View File

@ -14,6 +14,7 @@ import cats.effect.unsafe.implicits.global
import fs2.io.file.Path
import docspell.common._
import docspell.common.util.File
import docspell.convert._
import docspell.files.ExampleFiles
import docspell.logging.TestLoggingConfig

View File

@ -11,6 +11,7 @@ import fs2.Stream
import fs2.io.file.Path
import docspell.common._
import docspell.common.util.File
import docspell.logging.Logger
object Ocr {

View File

@ -11,6 +11,7 @@ import java.nio.file.Paths
import fs2.io.file.Path
import docspell.common._
import docspell.common.util.File
case class OcrConfig(
maxImageSize: Int,

View File

@ -0,0 +1,61 @@
/*
* Copyright 2020 Eike K. & Contributors
*
* SPDX-License-Identifier: AGPL-3.0-or-later
*/
package docspell.files
import cats.data.OptionT
import cats.effect.Sync
import cats.syntax.all._
import fs2.Stream
import fs2.io.file.{Files, Path}
import docspell.common.{MimeType, MimeTypeHint}
import io.circe.Encoder
import io.circe.syntax._
trait FileSupport {
implicit final class FileOps[F[_]: Files: Sync](self: Path) {
def detectMime: F[Option[MimeType]] =
Files[F].isReadable(self).flatMap { flag =>
OptionT
.whenF(flag) {
TikaMimetype
.detect(
Files[F].readAll(self),
MimeTypeHint.filename(self.fileName.toString)
)
}
.value
}
def asTextFile(alt: MimeType => F[Unit]): F[Option[Path]] =
OptionT(detectMime).flatMapF { mime =>
if (mime.matches(MimeType.text("plain"))) self.some.pure[F]
else alt(mime).as(None: Option[Path])
}.value
def readText: F[String] =
Files[F]
.readAll(self)
.through(fs2.text.utf8.decode)
.compile
.string
def readAll: Stream[F, Byte] =
Files[F].readAll(self)
def writeJson[A: Encoder](value: A): F[Unit] =
Stream
.emit(value.asJson.noSpaces)
.through(fs2.text.utf8.encode)
.through(Files[F].writeAll(self))
.compile
.drain
}
}
object FileSupport extends FileSupport

View File

@ -8,11 +8,12 @@ package docspell.files
import java.io.InputStream
import java.nio.charset.StandardCharsets
import java.nio.file.Paths
import java.util.zip.{ZipEntry, ZipInputStream, ZipOutputStream}
import cats.data.OptionT
import cats.effect._
import cats.implicits._
import fs2.io.file.{Files, Path}
import fs2.{Pipe, Stream}
import docspell.common.Binary
@ -27,16 +28,72 @@ object Zip {
): Pipe[F, (String, Stream[F, Byte]), Byte] =
in => zipJava(logger, chunkSize, in.through(deduplicate))
def unzipP[F[_]: Async](chunkSize: Int, glob: Glob): Pipe[F, Byte, Binary[F]] =
s => unzip[F](chunkSize, glob)(s)
def unzip[F[_]: Async](
chunkSize: Int,
glob: Glob
): Pipe[F, Byte, Binary[F]] =
s => unzipStream[F](chunkSize, glob)(s)
def unzip[F[_]: Async](chunkSize: Int, glob: Glob)(
def unzipStream[F[_]: Async](chunkSize: Int, glob: Glob)(
data: Stream[F, Byte]
): Stream[F, Binary[F]] =
data
.through(fs2.io.toInputStream[F])
.flatMap(in => unzipJava(in, chunkSize, glob))
def saveTo[F[_]: Async](
logger: Logger[F],
targetDir: Path,
moveUp: Boolean
): Pipe[F, Binary[F], Path] =
binaries =>
binaries
.filter(e => !e.name.endsWith("/"))
.evalMap { entry =>
val out = targetDir / entry.name
val createParent =
OptionT
.fromOption[F](out.parent)
.flatMapF(parent =>
Files[F]
.exists(parent)
.map(flag => Option.when(!flag)(parent))
)
.semiflatMap(p => Files[F].createDirectories(p))
.getOrElse(())
logger.trace(s"Unzip ${entry.name} -> $out") *>
createParent *>
entry.data.through(Files[F].writeAll(out)).compile.drain
}
.drain ++ Stream
.eval(if (moveUp) moveContentsUp(logger)(targetDir) else ().pure[F])
.as(targetDir)
private def moveContentsUp[F[_]: Sync: Files](logger: Logger[F])(dir: Path): F[Unit] =
Files[F]
.list(dir)
.take(2)
.compile
.toList
.flatMap {
case subdir :: Nil =>
Files[F].isDirectory(subdir).flatMap {
case false => ().pure[F]
case true =>
Files[F]
.list(subdir)
.filter(p => p != dir)
.evalTap(c => logger.trace(s"Move $c -> ${dir / c.fileName}"))
.evalMap(child => Files[F].move(child, dir / child.fileName))
.compile
.drain
}
case _ =>
().pure[F]
}
def unzipJava[F[_]: Async](
in: InputStream,
chunkSize: Int,
@ -55,7 +112,7 @@ object Zip {
.unNoneTerminate
.filter(ze => glob.matchFilenameOrPath(ze.getName()))
.map { ze =>
val name = Paths.get(ze.getName()).getFileName.toString
val name = ze.getName()
val data =
fs2.io.readInputStream[F]((zin: InputStream).pure[F], chunkSize, false)
Binary(name, data)

Binary file not shown.

Binary file not shown.

View File

@ -7,20 +7,25 @@
package docspell.files
import cats.effect._
import cats.effect.unsafe.implicits.global
import cats.implicits._
import fs2.io.file.{Files, Path}
import docspell.common.Glob
import docspell.logging.TestLoggingConfig
import munit._
class ZipTest extends FunSuite {
class ZipTest extends CatsEffectSuite with TestLoggingConfig {
val logger = docspell.logging.getLogger[IO]
val tempDir = ResourceFixture(
Files[IO].tempDirectory(Path("target").some, "zip-test-", None)
)
test("unzip") {
val zipFile = ExampleFiles.letters_zip.readURL[IO](8192)
val uncomp = zipFile.through(Zip.unzip(8192, Glob.all))
val unzip = zipFile.through(Zip.unzip(8192, Glob.all))
uncomp
unzip
.evalMap { entry =>
val x = entry.data.map(_ => 1).foldMonoid.compile.lastOrError
x.map { size =>
@ -35,6 +40,10 @@ class ZipTest extends FunSuite {
}
.compile
.drain
.unsafeRunSync()
}
tempDir.test("unzipTo directory tree") { _ =>
// val zipFile = ExampleFiles.zip_dirs_zip.readURL[IO](8192)
// zipFile.through(Zip.unzip(G))
}
}

View File

@ -780,4 +780,75 @@ Docpell Update Check
index-all-chunk = 10
}
}
addons {
# A directory to extract addons when running them. Everything in
# here will be cleared after each run.
working-dir = ${java.io.tmpdir}"/docspell-addons"
# A directory for addons to store data between runs. This is not
# cleared by Docspell and can get large depending on the addons
# executed.
#
# This directory is used as base. In it subdirectories are created
# per run configuration id.
cache-dir = ${java.io.tmpdir}"/docspell-addon-cache"
executor-config {
# Define a (comma or whitespace separated) list of runners that
# are responsible for executing an addon. This setting is
# compared to what is supported by addons. Possible values are:
#
# - nix-flake: use nix-flake runner if the addon supports it
# (this requires the nix package manager on the joex machine)
# - docker: use docker
# - trivial: use the trivial runner
#
# The first successful execution is used. This should list all
# runners the computer supports.
runner = "nix-flake, docker, trivial"
# systemd-nspawn can be used to run the program in a container.
# This is used by runners nix-flake and trivial.
nspawn = {
# If this is false, systemd-nspawn is not tried. When true, the
# addon is executed inside a lightweight container via
# systemd-nspawn.
enabled = false
# Path to sudo command. By default systemd-nspawn is executed
# via sudo - the user running joex must be allowed to do so NON
# INTERACTIVELY. If this is empty, then nspawn is tried to
# execute without sudo.
sudo-binary = "sudo"
# Path to the systemd-nspawn command.
nspawn-binary = "systemd-nspawn"
# Workaround, if multiple same named containers are run too fast
container-wait = "100 millis"
}
# The timeout for running an addon.
run-timeout = "15 minutes"
# Configure the nix flake runner.
nix-runner {
# Path to the nix command.
nix-binary = "nix"
# The timeout for building the package (running nix build).
build-timeout = "15 minutes"
}
# Configure the docker runner
docker-runner {
# Path to the docker command.
docker-binary = "docker"
# The timeout for building the package (running docker build).
build-timeout = "15 minutes"
}
}
}
}

View File

@ -12,6 +12,7 @@ import fs2.io.file.Path
import docspell.analysis.TextAnalysisConfig
import docspell.analysis.classifier.TextClassifierConfig
import docspell.backend.Config.Files
import docspell.backend.joex.AddonEnvConfig
import docspell.common._
import docspell.config.{FtsType, PgFtsConfig}
import docspell.convert.ConvertConfig
@ -43,7 +44,8 @@ case class Config(
files: Files,
mailDebug: Boolean,
fullTextSearch: Config.FullTextSearch,
updateCheck: UpdateCheckConfig
updateCheck: UpdateCheckConfig,
addons: AddonEnvConfig
) {
def pubSubConfig(headerValue: Ident): PubSubConfig =

View File

@ -145,6 +145,8 @@ object JoexAppImpl extends MailAddressCodec {
schedulerModule.scheduler,
schedulerModule.periodicScheduler
)
nodes <- ONode(store)
_ <- nodes.withRegistered(cfg.appId, NodeType.Joex, cfg.baseUrl, None)
appR <- Resource.make(app.init.map(_ => app))(_.initShutdown)
} yield appR

View File

@ -59,7 +59,7 @@ object JoexServer {
Router("pubsub" -> pubSub.receiveRoute)
},
"/api/info" -> InfoRoutes(cfg),
"/api/v1" -> JoexRoutes(joexApp)
"/api/v1" -> JoexRoutes(cfg, joexApp)
).orNotFound
// With Middlewares in place

Some files were not shown because too many files have changed in this diff Show More