Remove obsolete type capabilities

These are now detected by the new scala compiler and lead to compile
errors.
This commit is contained in:
Eike Kettner 2021-03-01 00:08:25 +01:00
parent 41bc33a73b
commit e6d9ce2c37
31 changed files with 40 additions and 42 deletions

View File

@ -696,8 +696,7 @@ def packageTools(logger: Logger, dir: File, version: String): Seq[File] = {
wx / "icons" / "logo-96.png" -> "icons/logo-96.png",
wx / "manifest.json" -> "manifest.json"
),
webext,
None
webext
)
val excludes = Seq(wx, target)
@ -714,8 +713,7 @@ def packageTools(logger: Logger, dir: File, version: String): Seq[File] = {
wx / "native/app_manifest.json" -> s"docspell-tools-${version}/firefox/native/app_manifest.json",
wx / "native/native.py" -> s"docspell-tools-${version}/firefox/native/native.py"
) ++ files,
archive,
None
archive
)
Seq(archive)

View File

@ -82,7 +82,7 @@ object TextAnalyser {
/** Provides the nlp pipeline based on the configuration. */
private object Nlp {
def apply[F[_]: Concurrent: Timer: BracketThrow](
def apply[F[_]: Concurrent: Timer](
cfg: TextAnalysisConfig.NlpConfig
): F[Input[F] => F[Vector[NerLabel]]] =
cfg.mode match {

View File

@ -33,7 +33,7 @@ import emil.javamail._
import org.http4s.client.Client
import org.http4s.client.blaze.BlazeClientBuilder
final class JoexAppImpl[F[_]: ConcurrentEffect: ContextShift: Timer](
final class JoexAppImpl[F[_]: ConcurrentEffect: Timer](
cfg: Config,
nodeOps: ONode[F],
store: Store[F],
@ -214,7 +214,7 @@ object JoexAppImpl {
appR <- Resource.make(app.init.map(_ => app))(_.shutdown)
} yield appR
private def createFtsClient[F[_]: ConcurrentEffect: ContextShift](
private def createFtsClient[F[_]: ConcurrentEffect](
cfg: Config
)(client: Client[F]): Resource[F, FtsClient[F]] =
if (cfg.fullTextSearch.enabled) SolrFtsClient(cfg.fullTextSearch.solr, client)

View File

@ -25,7 +25,7 @@ object MigrationTask {
)
)
def onCancel[F[_]: Sync]: Task[F, Unit, Unit] =
def onCancel[F[_]]: Task[F, Unit, Unit] =
Task.log[F, Unit](_.warn("Cancelling full-text-index task"))
def job[F[_]: Sync]: F[RJob] =

View File

@ -24,7 +24,7 @@ object ReIndexTask {
Task(ctx => clearData[F](ctx.args.collective).forContext(cfg, fts).run(ctx))
)
def onCancel[F[_]: Sync]: Task[F, Args, Unit] =
def onCancel[F[_]]: Task[F, Args, Unit] =
Task.log[F, Args](_.warn("Cancelling full-text re-index task"))
private def clearData[F[_]: ConcurrentEffect](collective: Option[Ident]): FtsWork[F] =

View File

@ -58,7 +58,7 @@ object CheckNodesTask {
} yield res.isRight
}
def removeNodes[F[_]: Sync](
def removeNodes[F[_]](
ctx: Context[F, _],
cfg: HouseKeepingConfig.CheckNodes
): F[Int] =

View File

@ -23,7 +23,7 @@ object HouseKeepingTask {
.flatMap(_ => CleanupJobsTask(cfg.houseKeeping.cleanupJobs))
.flatMap(_ => CheckNodesTask(cfg.houseKeeping.checkNodes))
def onCancel[F[_]: Sync]: Task[F, Unit, Unit] =
def onCancel[F[_]]: Task[F, Unit, Unit] =
Task.log[F, Unit](_.warn("Cancelling house-keeping task"))
def periodicTask[F[_]: Sync](ce: CalEvent): F[RPeriodicTask] =

View File

@ -17,7 +17,7 @@ object LearnClassifierTask {
type Args = LearnClassifierArgs
def onCancel[F[_]: Sync]: Task[F, Args, Unit] =
def onCancel[F[_]]: Task[F, Args, Unit] =
Task.log(_.warn("Cancelling learn-classifier task"))
def apply[F[_]: Sync: ContextShift](
@ -100,7 +100,7 @@ object LearnClassifierTask {
else
OptionT.none
private def logInactiveWarning[F[_]: Sync](logger: Logger[F]): F[Unit] =
private def logInactiveWarning[F[_]](logger: Logger[F]): F[Unit] =
logger.warn(
"Auto-tagging is disabled. Check joex config and the collective settings."
)

View File

@ -15,7 +15,7 @@ import emil.{MimeType => _, _}
object ReadMail {
def readBytesP[F[_]: ConcurrentEffect: ContextShift](
def readBytesP[F[_]: ConcurrentEffect](
logger: Logger[F],
glob: Glob
): Pipe[F, Byte, Binary[F]] =
@ -26,7 +26,7 @@ object ReadMail {
Stream.eval(logger.debug(s"Converting e-mail file...")) >>
s.through(Mail.readBytes[F])
def mailToEntries[F[_]: ConcurrentEffect: ContextShift](
def mailToEntries[F[_]: ConcurrentEffect](
logger: Logger[F],
glob: Glob
)(mail: Mail[F]): Stream[F, Binary[F]] = {

View File

@ -40,7 +40,7 @@ object NotifyDueItemsTask {
} yield ()
}
def onCancel[F[_]: Sync]: Task[F, NotifyDueItemsArgs, Unit] =
def onCancel[F[_]]: Task[F, NotifyDueItemsArgs, Unit] =
Task.log(_.warn("Cancelling notify-due-items task"))
def getMailSettings[F[_]: Sync](ctx: Context[F, Args]): F[RUserEmail] =

View File

@ -28,7 +28,7 @@ object AllPageCountTask {
} yield ()
}
def onCancel[F[_]: Sync]: Task[F, Args, Unit] =
def onCancel[F[_]]: Task[F, Args, Unit] =
Task.log(_.warn("Cancelling all-previews task"))
def submitConversionJobs[F[_]: Sync](

View File

@ -30,7 +30,7 @@ object MakePageCountTask {
} yield ()
}
def onCancel[F[_]: Sync]: Task[F, Args, Unit] =
def onCancel[F[_]]: Task[F, Args, Unit] =
Task.log(_.warn("Cancelling make-page-count task"))
private def generatePageCount[F[_]: Sync](
@ -45,7 +45,7 @@ object MakePageCountTask {
)
} yield ()
private def pageCountExists[F[_]: Sync](ctx: Context[F, Args]): F[Boolean] =
private def pageCountExists[F[_]](ctx: Context[F, Args]): F[Boolean] =
ctx.store.transact(
RAttachmentMeta
.findPageCountById(ctx.args.attachment)

View File

@ -28,7 +28,7 @@ object ConvertAllPdfTask {
} yield ()
}
def onCancel[F[_]: Sync]: Task[F, Args, Unit] =
def onCancel[F[_]]: Task[F, Args, Unit] =
Task.log(_.warn("Cancelling convert-old-pdf task"))
def submitConversionJobs[F[_]: Sync](

View File

@ -44,7 +44,7 @@ object PdfConvTask {
} yield ()
}
def onCancel[F[_]: Sync]: Task[F, Args, Unit] =
def onCancel[F[_]]: Task[F, Args, Unit] =
Task.log(_.warn("Cancelling pdfconv task"))
// --- Helper

View File

@ -28,7 +28,7 @@ object AllPreviewsTask {
} yield ()
}
def onCancel[F[_]: Sync]: Task[F, Args, Unit] =
def onCancel[F[_]]: Task[F, Args, Unit] =
Task.log(_.warn("Cancelling all-previews task"))
def submitConversionJobs[F[_]: Sync](

View File

@ -34,7 +34,7 @@ object MakePreviewTask {
} yield ()
}
def onCancel[F[_]: Sync]: Task[F, Args, Unit] =
def onCancel[F[_]]: Task[F, Args, Unit] =
Task.log(_.warn("Cancelling make-preview task"))
private def generatePreview[F[_]: Sync](

View File

@ -22,7 +22,7 @@ import bitpeace.{Mimetype, RangeDef}
*/
object AttachmentPageCount {
def apply[F[_]: Sync: ContextShift]()(
def apply[F[_]: Sync]()(
item: ItemData
): Task[F, ProcessItemArgs, ItemData] =
Task { ctx =>

View File

@ -24,7 +24,7 @@ import bitpeace.{Mimetype, MimetypeHint, RangeDef}
*/
object AttachmentPreview {
def apply[F[_]: Sync: ContextShift](cfg: ConvertConfig, pcfg: PreviewConfig)(
def apply[F[_]: Sync](cfg: ConvertConfig, pcfg: PreviewConfig)(
item: ItemData
): Task[F, ProcessItemArgs, ItemData] =
Task { ctx =>

View File

@ -61,7 +61,7 @@ object ConvertPdf {
}
def isConverted[F[_]: Sync](ctx: Context[F, ProcessItemArgs])(
def isConverted[F[_]](ctx: Context[F, ProcessItemArgs])(
ra: RAttachment
): F[Boolean] =
ctx.store.transact(RAttachmentSource.isConverted(ra.id))

View File

@ -113,7 +113,7 @@ object CreateItem {
)
}
def insertAttachment[F[_]: Sync](ctx: Context[F, _])(ra: RAttachment): F[Int] = {
def insertAttachment[F[_]](ctx: Context[F, _])(ra: RAttachment): F[Int] = {
val rs = RAttachmentSource.of(ra)
ctx.store.transact(for {
n <- RAttachment.insert(ra)

View File

@ -53,7 +53,7 @@ object DuplicateCheck {
else ().pure[F]
}
private def findDuplicates[F[_]: Sync](
private def findDuplicates[F[_]](
ctx: Context[F, Args]
): F[Vector[FileMetaDupes]] =
ctx.store.transact(for {

View File

@ -150,7 +150,7 @@ object ExtractArchive {
.lastOrError
}
def extractMail[F[_]: ConcurrentEffect: ContextShift](
def extractMail[F[_]: ConcurrentEffect](
ctx: Context[F, ProcessItemArgs],
archive: Option[RAttachmentArchive]
)(ra: RAttachment, pos: Int): F[Extracted] = {
@ -225,7 +225,7 @@ object ExtractArchive {
} yield n1 + n2
}
def storeArchive[F[_]: Sync](ctx: Context[F, _])(aa: RAttachmentArchive): F[Int] =
def storeArchive[F[_]](ctx: Context[F, _])(aa: RAttachmentArchive): F[Int] =
ctx.store.transact(RAttachmentArchive.insert(aa))
case class Extracted(

View File

@ -18,7 +18,7 @@ import docspell.store.records.RItem
object ItemHandler {
type Args = ProcessItemArgs
def onCancel[F[_]: Sync: ContextShift]: Task[F, Args, Unit] =
def onCancel[F[_]: Sync]: Task[F, Args, Unit] =
logWarn("Now cancelling.").flatMap(_ =>
markItemCreated.flatMap {
case true =>
@ -105,7 +105,7 @@ object ItemHandler {
)
}
private def deleteByFileIds[F[_]: Sync: ContextShift]: Task[F, Args, Unit] =
private def deleteByFileIds[F[_]: Sync]: Task[F, Args, Unit] =
Task { ctx =>
val states = ItemState.invalidStates
for {

View File

@ -29,7 +29,7 @@ object ReProcessItem {
.flatMap(safeProcess[F](cfg, fts, analyser, regexNer))
.map(_ => ())
def onCancel[F[_]: Sync: ContextShift]: Task[F, Args, Unit] =
def onCancel[F[_]]: Task[F, Args, Unit] =
logWarn("Now cancelling re-processing.")
// --- Helpers

View File

@ -119,7 +119,7 @@ object TextExtraction {
)
} yield (meta, tags)
def extractText[F[_]: Sync: ContextShift](
def extractText[F[_]: Sync](
ctx: Context[F, _],
extr: Extraction[F],
lang: Language

View File

@ -46,7 +46,7 @@ object ScanMailboxTask {
} yield ()
}
def onCancel[F[_]: Sync]: Task[F, ScanMailboxArgs, Unit] =
def onCancel[F[_]]: Task[F, ScanMailboxArgs, Unit] =
Task.log(_.warn("Cancelling scan-mailbox task"))
def getMailSettings[F[_]: Sync](ctx: Context[F, Args]): F[RUserImap] =

View File

@ -31,7 +31,7 @@ trait PeriodicScheduler[F[_]] {
object PeriodicScheduler {
def create[F[_]: ConcurrentEffect: ContextShift](
def create[F[_]: ConcurrentEffect](
cfg: PeriodicSchedulerConfig,
sch: Scheduler[F],
queue: JobQueue[F],

View File

@ -15,7 +15,7 @@ import docspell.store.records.RPeriodicTask
import com.github.eikek.fs2calev._
import org.log4s.getLogger
final class PeriodicSchedulerImpl[F[_]: ConcurrentEffect: ContextShift](
final class PeriodicSchedulerImpl[F[_]: ConcurrentEffect](
val config: PeriodicSchedulerConfig,
sch: Scheduler[F],
queue: JobQueue[F],

View File

@ -12,7 +12,7 @@ import docspell.ftssolr.SolrFtsClient
import org.http4s.client.Client
final class RestAppImpl[F[_]: Sync](val config: Config, val backend: BackendApp[F])
final class RestAppImpl[F[_]](val config: Config, val backend: BackendApp[F])
extends RestApp[F] {
def init: F[Unit] =
@ -38,7 +38,7 @@ object RestAppImpl {
appR <- Resource.make(app.init.map(_ => app))(_.shutdown)
} yield appR
private def createFtsClient[F[_]: ConcurrentEffect: ContextShift](
private def createFtsClient[F[_]: ConcurrentEffect](
cfg: Config
)(client: Client[F]): Resource[F, FtsClient[F]] =
if (cfg.fullTextSearch.enabled) SolrFtsClient(cfg.fullTextSearch.solr, client)

View File

@ -606,7 +606,7 @@ trait Conversions {
)
})
def changeSource[F[_]: Sync](s: Source, coll: Ident): RSource =
def changeSource[F[_]](s: Source, coll: Ident): RSource =
RSource(
s.id,
coll,

View File

@ -1,7 +1,7 @@
package docspell.store.queries
import cats.data.NonEmptyList
import cats.effect.Effect
import cats.effect._
import cats.implicits._
import fs2.Stream
@ -197,7 +197,7 @@ object QJob {
def runningToWaiting[F[_]: Effect](workerId: Ident, store: Store[F]): F[Unit] =
store.transact(RJob.setRunningToWaiting(workerId)).map(_ => ())
def findAll[F[_]: Effect](ids: Seq[Ident], store: Store[F]): F[Vector[RJob]] =
def findAll[F[_]](ids: Seq[Ident], store: Store[F]): F[Vector[RJob]] =
store.transact(RJob.findFromIds(ids))
def queueStateSnapshot(