mirror of
https://github.com/TheAnachronism/docspell.git
synced 2025-04-04 18:39:33 +00:00
Remove obsolete type capabilities
These are now detected by the new scala compiler and lead to compile errors.
This commit is contained in:
parent
41bc33a73b
commit
e6d9ce2c37
@ -696,8 +696,7 @@ def packageTools(logger: Logger, dir: File, version: String): Seq[File] = {
|
||||
wx / "icons" / "logo-96.png" -> "icons/logo-96.png",
|
||||
wx / "manifest.json" -> "manifest.json"
|
||||
),
|
||||
webext,
|
||||
None
|
||||
webext
|
||||
)
|
||||
|
||||
val excludes = Seq(wx, target)
|
||||
@ -714,8 +713,7 @@ def packageTools(logger: Logger, dir: File, version: String): Seq[File] = {
|
||||
wx / "native/app_manifest.json" -> s"docspell-tools-${version}/firefox/native/app_manifest.json",
|
||||
wx / "native/native.py" -> s"docspell-tools-${version}/firefox/native/native.py"
|
||||
) ++ files,
|
||||
archive,
|
||||
None
|
||||
archive
|
||||
)
|
||||
|
||||
Seq(archive)
|
||||
|
@ -82,7 +82,7 @@ object TextAnalyser {
|
||||
|
||||
/** Provides the nlp pipeline based on the configuration. */
|
||||
private object Nlp {
|
||||
def apply[F[_]: Concurrent: Timer: BracketThrow](
|
||||
def apply[F[_]: Concurrent: Timer](
|
||||
cfg: TextAnalysisConfig.NlpConfig
|
||||
): F[Input[F] => F[Vector[NerLabel]]] =
|
||||
cfg.mode match {
|
||||
|
@ -33,7 +33,7 @@ import emil.javamail._
|
||||
import org.http4s.client.Client
|
||||
import org.http4s.client.blaze.BlazeClientBuilder
|
||||
|
||||
final class JoexAppImpl[F[_]: ConcurrentEffect: ContextShift: Timer](
|
||||
final class JoexAppImpl[F[_]: ConcurrentEffect: Timer](
|
||||
cfg: Config,
|
||||
nodeOps: ONode[F],
|
||||
store: Store[F],
|
||||
@ -214,7 +214,7 @@ object JoexAppImpl {
|
||||
appR <- Resource.make(app.init.map(_ => app))(_.shutdown)
|
||||
} yield appR
|
||||
|
||||
private def createFtsClient[F[_]: ConcurrentEffect: ContextShift](
|
||||
private def createFtsClient[F[_]: ConcurrentEffect](
|
||||
cfg: Config
|
||||
)(client: Client[F]): Resource[F, FtsClient[F]] =
|
||||
if (cfg.fullTextSearch.enabled) SolrFtsClient(cfg.fullTextSearch.solr, client)
|
||||
|
@ -25,7 +25,7 @@ object MigrationTask {
|
||||
)
|
||||
)
|
||||
|
||||
def onCancel[F[_]: Sync]: Task[F, Unit, Unit] =
|
||||
def onCancel[F[_]]: Task[F, Unit, Unit] =
|
||||
Task.log[F, Unit](_.warn("Cancelling full-text-index task"))
|
||||
|
||||
def job[F[_]: Sync]: F[RJob] =
|
||||
|
@ -24,7 +24,7 @@ object ReIndexTask {
|
||||
Task(ctx => clearData[F](ctx.args.collective).forContext(cfg, fts).run(ctx))
|
||||
)
|
||||
|
||||
def onCancel[F[_]: Sync]: Task[F, Args, Unit] =
|
||||
def onCancel[F[_]]: Task[F, Args, Unit] =
|
||||
Task.log[F, Args](_.warn("Cancelling full-text re-index task"))
|
||||
|
||||
private def clearData[F[_]: ConcurrentEffect](collective: Option[Ident]): FtsWork[F] =
|
||||
|
@ -58,7 +58,7 @@ object CheckNodesTask {
|
||||
} yield res.isRight
|
||||
}
|
||||
|
||||
def removeNodes[F[_]: Sync](
|
||||
def removeNodes[F[_]](
|
||||
ctx: Context[F, _],
|
||||
cfg: HouseKeepingConfig.CheckNodes
|
||||
): F[Int] =
|
||||
|
@ -23,7 +23,7 @@ object HouseKeepingTask {
|
||||
.flatMap(_ => CleanupJobsTask(cfg.houseKeeping.cleanupJobs))
|
||||
.flatMap(_ => CheckNodesTask(cfg.houseKeeping.checkNodes))
|
||||
|
||||
def onCancel[F[_]: Sync]: Task[F, Unit, Unit] =
|
||||
def onCancel[F[_]]: Task[F, Unit, Unit] =
|
||||
Task.log[F, Unit](_.warn("Cancelling house-keeping task"))
|
||||
|
||||
def periodicTask[F[_]: Sync](ce: CalEvent): F[RPeriodicTask] =
|
||||
|
@ -17,7 +17,7 @@ object LearnClassifierTask {
|
||||
|
||||
type Args = LearnClassifierArgs
|
||||
|
||||
def onCancel[F[_]: Sync]: Task[F, Args, Unit] =
|
||||
def onCancel[F[_]]: Task[F, Args, Unit] =
|
||||
Task.log(_.warn("Cancelling learn-classifier task"))
|
||||
|
||||
def apply[F[_]: Sync: ContextShift](
|
||||
@ -100,7 +100,7 @@ object LearnClassifierTask {
|
||||
else
|
||||
OptionT.none
|
||||
|
||||
private def logInactiveWarning[F[_]: Sync](logger: Logger[F]): F[Unit] =
|
||||
private def logInactiveWarning[F[_]](logger: Logger[F]): F[Unit] =
|
||||
logger.warn(
|
||||
"Auto-tagging is disabled. Check joex config and the collective settings."
|
||||
)
|
||||
|
@ -15,7 +15,7 @@ import emil.{MimeType => _, _}
|
||||
|
||||
object ReadMail {
|
||||
|
||||
def readBytesP[F[_]: ConcurrentEffect: ContextShift](
|
||||
def readBytesP[F[_]: ConcurrentEffect](
|
||||
logger: Logger[F],
|
||||
glob: Glob
|
||||
): Pipe[F, Byte, Binary[F]] =
|
||||
@ -26,7 +26,7 @@ object ReadMail {
|
||||
Stream.eval(logger.debug(s"Converting e-mail file...")) >>
|
||||
s.through(Mail.readBytes[F])
|
||||
|
||||
def mailToEntries[F[_]: ConcurrentEffect: ContextShift](
|
||||
def mailToEntries[F[_]: ConcurrentEffect](
|
||||
logger: Logger[F],
|
||||
glob: Glob
|
||||
)(mail: Mail[F]): Stream[F, Binary[F]] = {
|
||||
|
@ -40,7 +40,7 @@ object NotifyDueItemsTask {
|
||||
} yield ()
|
||||
}
|
||||
|
||||
def onCancel[F[_]: Sync]: Task[F, NotifyDueItemsArgs, Unit] =
|
||||
def onCancel[F[_]]: Task[F, NotifyDueItemsArgs, Unit] =
|
||||
Task.log(_.warn("Cancelling notify-due-items task"))
|
||||
|
||||
def getMailSettings[F[_]: Sync](ctx: Context[F, Args]): F[RUserEmail] =
|
||||
|
@ -28,7 +28,7 @@ object AllPageCountTask {
|
||||
} yield ()
|
||||
}
|
||||
|
||||
def onCancel[F[_]: Sync]: Task[F, Args, Unit] =
|
||||
def onCancel[F[_]]: Task[F, Args, Unit] =
|
||||
Task.log(_.warn("Cancelling all-previews task"))
|
||||
|
||||
def submitConversionJobs[F[_]: Sync](
|
||||
|
@ -30,7 +30,7 @@ object MakePageCountTask {
|
||||
} yield ()
|
||||
}
|
||||
|
||||
def onCancel[F[_]: Sync]: Task[F, Args, Unit] =
|
||||
def onCancel[F[_]]: Task[F, Args, Unit] =
|
||||
Task.log(_.warn("Cancelling make-page-count task"))
|
||||
|
||||
private def generatePageCount[F[_]: Sync](
|
||||
@ -45,7 +45,7 @@ object MakePageCountTask {
|
||||
)
|
||||
} yield ()
|
||||
|
||||
private def pageCountExists[F[_]: Sync](ctx: Context[F, Args]): F[Boolean] =
|
||||
private def pageCountExists[F[_]](ctx: Context[F, Args]): F[Boolean] =
|
||||
ctx.store.transact(
|
||||
RAttachmentMeta
|
||||
.findPageCountById(ctx.args.attachment)
|
||||
|
@ -28,7 +28,7 @@ object ConvertAllPdfTask {
|
||||
} yield ()
|
||||
}
|
||||
|
||||
def onCancel[F[_]: Sync]: Task[F, Args, Unit] =
|
||||
def onCancel[F[_]]: Task[F, Args, Unit] =
|
||||
Task.log(_.warn("Cancelling convert-old-pdf task"))
|
||||
|
||||
def submitConversionJobs[F[_]: Sync](
|
||||
|
@ -44,7 +44,7 @@ object PdfConvTask {
|
||||
} yield ()
|
||||
}
|
||||
|
||||
def onCancel[F[_]: Sync]: Task[F, Args, Unit] =
|
||||
def onCancel[F[_]]: Task[F, Args, Unit] =
|
||||
Task.log(_.warn("Cancelling pdfconv task"))
|
||||
|
||||
// --- Helper
|
||||
|
@ -28,7 +28,7 @@ object AllPreviewsTask {
|
||||
} yield ()
|
||||
}
|
||||
|
||||
def onCancel[F[_]: Sync]: Task[F, Args, Unit] =
|
||||
def onCancel[F[_]]: Task[F, Args, Unit] =
|
||||
Task.log(_.warn("Cancelling all-previews task"))
|
||||
|
||||
def submitConversionJobs[F[_]: Sync](
|
||||
|
@ -34,7 +34,7 @@ object MakePreviewTask {
|
||||
} yield ()
|
||||
}
|
||||
|
||||
def onCancel[F[_]: Sync]: Task[F, Args, Unit] =
|
||||
def onCancel[F[_]]: Task[F, Args, Unit] =
|
||||
Task.log(_.warn("Cancelling make-preview task"))
|
||||
|
||||
private def generatePreview[F[_]: Sync](
|
||||
|
@ -22,7 +22,7 @@ import bitpeace.{Mimetype, RangeDef}
|
||||
*/
|
||||
object AttachmentPageCount {
|
||||
|
||||
def apply[F[_]: Sync: ContextShift]()(
|
||||
def apply[F[_]: Sync]()(
|
||||
item: ItemData
|
||||
): Task[F, ProcessItemArgs, ItemData] =
|
||||
Task { ctx =>
|
||||
|
@ -24,7 +24,7 @@ import bitpeace.{Mimetype, MimetypeHint, RangeDef}
|
||||
*/
|
||||
object AttachmentPreview {
|
||||
|
||||
def apply[F[_]: Sync: ContextShift](cfg: ConvertConfig, pcfg: PreviewConfig)(
|
||||
def apply[F[_]: Sync](cfg: ConvertConfig, pcfg: PreviewConfig)(
|
||||
item: ItemData
|
||||
): Task[F, ProcessItemArgs, ItemData] =
|
||||
Task { ctx =>
|
||||
|
@ -61,7 +61,7 @@ object ConvertPdf {
|
||||
|
||||
}
|
||||
|
||||
def isConverted[F[_]: Sync](ctx: Context[F, ProcessItemArgs])(
|
||||
def isConverted[F[_]](ctx: Context[F, ProcessItemArgs])(
|
||||
ra: RAttachment
|
||||
): F[Boolean] =
|
||||
ctx.store.transact(RAttachmentSource.isConverted(ra.id))
|
||||
|
@ -113,7 +113,7 @@ object CreateItem {
|
||||
)
|
||||
}
|
||||
|
||||
def insertAttachment[F[_]: Sync](ctx: Context[F, _])(ra: RAttachment): F[Int] = {
|
||||
def insertAttachment[F[_]](ctx: Context[F, _])(ra: RAttachment): F[Int] = {
|
||||
val rs = RAttachmentSource.of(ra)
|
||||
ctx.store.transact(for {
|
||||
n <- RAttachment.insert(ra)
|
||||
|
@ -53,7 +53,7 @@ object DuplicateCheck {
|
||||
else ().pure[F]
|
||||
}
|
||||
|
||||
private def findDuplicates[F[_]: Sync](
|
||||
private def findDuplicates[F[_]](
|
||||
ctx: Context[F, Args]
|
||||
): F[Vector[FileMetaDupes]] =
|
||||
ctx.store.transact(for {
|
||||
|
@ -150,7 +150,7 @@ object ExtractArchive {
|
||||
.lastOrError
|
||||
}
|
||||
|
||||
def extractMail[F[_]: ConcurrentEffect: ContextShift](
|
||||
def extractMail[F[_]: ConcurrentEffect](
|
||||
ctx: Context[F, ProcessItemArgs],
|
||||
archive: Option[RAttachmentArchive]
|
||||
)(ra: RAttachment, pos: Int): F[Extracted] = {
|
||||
@ -225,7 +225,7 @@ object ExtractArchive {
|
||||
} yield n1 + n2
|
||||
}
|
||||
|
||||
def storeArchive[F[_]: Sync](ctx: Context[F, _])(aa: RAttachmentArchive): F[Int] =
|
||||
def storeArchive[F[_]](ctx: Context[F, _])(aa: RAttachmentArchive): F[Int] =
|
||||
ctx.store.transact(RAttachmentArchive.insert(aa))
|
||||
|
||||
case class Extracted(
|
||||
|
@ -18,7 +18,7 @@ import docspell.store.records.RItem
|
||||
object ItemHandler {
|
||||
type Args = ProcessItemArgs
|
||||
|
||||
def onCancel[F[_]: Sync: ContextShift]: Task[F, Args, Unit] =
|
||||
def onCancel[F[_]: Sync]: Task[F, Args, Unit] =
|
||||
logWarn("Now cancelling.").flatMap(_ =>
|
||||
markItemCreated.flatMap {
|
||||
case true =>
|
||||
@ -105,7 +105,7 @@ object ItemHandler {
|
||||
)
|
||||
}
|
||||
|
||||
private def deleteByFileIds[F[_]: Sync: ContextShift]: Task[F, Args, Unit] =
|
||||
private def deleteByFileIds[F[_]: Sync]: Task[F, Args, Unit] =
|
||||
Task { ctx =>
|
||||
val states = ItemState.invalidStates
|
||||
for {
|
||||
|
@ -29,7 +29,7 @@ object ReProcessItem {
|
||||
.flatMap(safeProcess[F](cfg, fts, analyser, regexNer))
|
||||
.map(_ => ())
|
||||
|
||||
def onCancel[F[_]: Sync: ContextShift]: Task[F, Args, Unit] =
|
||||
def onCancel[F[_]]: Task[F, Args, Unit] =
|
||||
logWarn("Now cancelling re-processing.")
|
||||
|
||||
// --- Helpers
|
||||
|
@ -119,7 +119,7 @@ object TextExtraction {
|
||||
)
|
||||
} yield (meta, tags)
|
||||
|
||||
def extractText[F[_]: Sync: ContextShift](
|
||||
def extractText[F[_]: Sync](
|
||||
ctx: Context[F, _],
|
||||
extr: Extraction[F],
|
||||
lang: Language
|
||||
|
@ -46,7 +46,7 @@ object ScanMailboxTask {
|
||||
} yield ()
|
||||
}
|
||||
|
||||
def onCancel[F[_]: Sync]: Task[F, ScanMailboxArgs, Unit] =
|
||||
def onCancel[F[_]]: Task[F, ScanMailboxArgs, Unit] =
|
||||
Task.log(_.warn("Cancelling scan-mailbox task"))
|
||||
|
||||
def getMailSettings[F[_]: Sync](ctx: Context[F, Args]): F[RUserImap] =
|
||||
|
@ -31,7 +31,7 @@ trait PeriodicScheduler[F[_]] {
|
||||
|
||||
object PeriodicScheduler {
|
||||
|
||||
def create[F[_]: ConcurrentEffect: ContextShift](
|
||||
def create[F[_]: ConcurrentEffect](
|
||||
cfg: PeriodicSchedulerConfig,
|
||||
sch: Scheduler[F],
|
||||
queue: JobQueue[F],
|
||||
|
@ -15,7 +15,7 @@ import docspell.store.records.RPeriodicTask
|
||||
import com.github.eikek.fs2calev._
|
||||
import org.log4s.getLogger
|
||||
|
||||
final class PeriodicSchedulerImpl[F[_]: ConcurrentEffect: ContextShift](
|
||||
final class PeriodicSchedulerImpl[F[_]: ConcurrentEffect](
|
||||
val config: PeriodicSchedulerConfig,
|
||||
sch: Scheduler[F],
|
||||
queue: JobQueue[F],
|
||||
|
@ -12,7 +12,7 @@ import docspell.ftssolr.SolrFtsClient
|
||||
|
||||
import org.http4s.client.Client
|
||||
|
||||
final class RestAppImpl[F[_]: Sync](val config: Config, val backend: BackendApp[F])
|
||||
final class RestAppImpl[F[_]](val config: Config, val backend: BackendApp[F])
|
||||
extends RestApp[F] {
|
||||
|
||||
def init: F[Unit] =
|
||||
@ -38,7 +38,7 @@ object RestAppImpl {
|
||||
appR <- Resource.make(app.init.map(_ => app))(_.shutdown)
|
||||
} yield appR
|
||||
|
||||
private def createFtsClient[F[_]: ConcurrentEffect: ContextShift](
|
||||
private def createFtsClient[F[_]: ConcurrentEffect](
|
||||
cfg: Config
|
||||
)(client: Client[F]): Resource[F, FtsClient[F]] =
|
||||
if (cfg.fullTextSearch.enabled) SolrFtsClient(cfg.fullTextSearch.solr, client)
|
||||
|
@ -606,7 +606,7 @@ trait Conversions {
|
||||
)
|
||||
})
|
||||
|
||||
def changeSource[F[_]: Sync](s: Source, coll: Ident): RSource =
|
||||
def changeSource[F[_]](s: Source, coll: Ident): RSource =
|
||||
RSource(
|
||||
s.id,
|
||||
coll,
|
||||
|
@ -1,7 +1,7 @@
|
||||
package docspell.store.queries
|
||||
|
||||
import cats.data.NonEmptyList
|
||||
import cats.effect.Effect
|
||||
import cats.effect._
|
||||
import cats.implicits._
|
||||
import fs2.Stream
|
||||
|
||||
@ -197,7 +197,7 @@ object QJob {
|
||||
def runningToWaiting[F[_]: Effect](workerId: Ident, store: Store[F]): F[Unit] =
|
||||
store.transact(RJob.setRunningToWaiting(workerId)).map(_ => ())
|
||||
|
||||
def findAll[F[_]: Effect](ids: Seq[Ident], store: Store[F]): F[Vector[RJob]] =
|
||||
def findAll[F[_]](ids: Seq[Ident], store: Store[F]): F[Vector[RJob]] =
|
||||
store.transact(RJob.findFromIds(ids))
|
||||
|
||||
def queueStateSnapshot(
|
||||
|
Loading…
x
Reference in New Issue
Block a user