mirror of
https://github.com/TheAnachronism/docspell.git
synced 2025-04-05 10:59:33 +00:00
Merge pull request #1010 from eikek/fixup/347-file-exists
Fixup for deleting items
This commit is contained in:
commit
5782166273
@ -11,6 +11,7 @@ import scala.concurrent.ExecutionContext
|
|||||||
import cats.effect._
|
import cats.effect._
|
||||||
|
|
||||||
import docspell.backend.auth.Login
|
import docspell.backend.auth.Login
|
||||||
|
import docspell.backend.fulltext.CreateIndex
|
||||||
import docspell.backend.ops._
|
import docspell.backend.ops._
|
||||||
import docspell.backend.signup.OSignup
|
import docspell.backend.signup.OSignup
|
||||||
import docspell.ftsclient.FtsClient
|
import docspell.ftsclient.FtsClient
|
||||||
@ -69,7 +70,8 @@ object BackendApp {
|
|||||||
uploadImpl <- OUpload(store, queue, cfg.files, joexImpl)
|
uploadImpl <- OUpload(store, queue, cfg.files, joexImpl)
|
||||||
nodeImpl <- ONode(store)
|
nodeImpl <- ONode(store)
|
||||||
jobImpl <- OJob(store, joexImpl)
|
jobImpl <- OJob(store, joexImpl)
|
||||||
itemImpl <- OItem(store, ftsClient, queue, joexImpl)
|
createIndex <- CreateIndex.resource(ftsClient, store)
|
||||||
|
itemImpl <- OItem(store, ftsClient, createIndex, queue, joexImpl)
|
||||||
itemSearchImpl <- OItemSearch(store)
|
itemSearchImpl <- OItemSearch(store)
|
||||||
fulltextImpl <- OFulltext(itemSearchImpl, ftsClient, store, queue, joexImpl)
|
fulltextImpl <- OFulltext(itemSearchImpl, ftsClient, store, queue, joexImpl)
|
||||||
javaEmil =
|
javaEmil =
|
||||||
|
@ -0,0 +1,71 @@
|
|||||||
|
/*
|
||||||
|
* Copyright 2020 Docspell Contributors
|
||||||
|
*
|
||||||
|
* SPDX-License-Identifier: GPL-3.0-or-later
|
||||||
|
*/
|
||||||
|
|
||||||
|
package docspell.backend.fulltext
|
||||||
|
|
||||||
|
import cats.data.NonEmptyList
|
||||||
|
import cats.effect._
|
||||||
|
|
||||||
|
import docspell.common._
|
||||||
|
import docspell.ftsclient.FtsClient
|
||||||
|
import docspell.ftsclient.TextData
|
||||||
|
import docspell.store.Store
|
||||||
|
import docspell.store.queries.QAttachment
|
||||||
|
import docspell.store.queries.QItem
|
||||||
|
|
||||||
|
trait CreateIndex[F[_]] {
|
||||||
|
|
||||||
|
/** Low-level function to re-index data. It is not submitted as a job,
|
||||||
|
* but invoked on the current machine.
|
||||||
|
*/
|
||||||
|
def reIndexData(
|
||||||
|
logger: Logger[F],
|
||||||
|
collective: Option[Ident],
|
||||||
|
itemIds: Option[NonEmptyList[Ident]],
|
||||||
|
chunkSize: Int
|
||||||
|
): F[Unit]
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
object CreateIndex {
|
||||||
|
|
||||||
|
def resource[F[_]](fts: FtsClient[F], store: Store[F]): Resource[F, CreateIndex[F]] =
|
||||||
|
Resource.pure(apply(fts, store))
|
||||||
|
|
||||||
|
def apply[F[_]](fts: FtsClient[F], store: Store[F]): CreateIndex[F] =
|
||||||
|
new CreateIndex[F] {
|
||||||
|
def reIndexData(
|
||||||
|
logger: Logger[F],
|
||||||
|
collective: Option[Ident],
|
||||||
|
itemIds: Option[NonEmptyList[Ident]],
|
||||||
|
chunkSize: Int
|
||||||
|
): F[Unit] = {
|
||||||
|
val attachs = store
|
||||||
|
.transact(QAttachment.allAttachmentMetaAndName(collective, itemIds, chunkSize))
|
||||||
|
.map(caa =>
|
||||||
|
TextData
|
||||||
|
.attachment(
|
||||||
|
caa.item,
|
||||||
|
caa.id,
|
||||||
|
caa.collective,
|
||||||
|
caa.folder,
|
||||||
|
caa.lang,
|
||||||
|
caa.name,
|
||||||
|
caa.content
|
||||||
|
)
|
||||||
|
)
|
||||||
|
|
||||||
|
val items = store
|
||||||
|
.transact(QItem.allNameAndNotes(collective, itemIds, chunkSize))
|
||||||
|
.map(nn =>
|
||||||
|
TextData.item(nn.id, nn.collective, nn.folder, Option(nn.name), nn.notes)
|
||||||
|
)
|
||||||
|
|
||||||
|
fts.indexData(logger, attachs ++ items)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
@ -11,6 +11,7 @@ import cats.effect.{Async, Resource}
|
|||||||
import cats.implicits._
|
import cats.implicits._
|
||||||
|
|
||||||
import docspell.backend.JobFactory
|
import docspell.backend.JobFactory
|
||||||
|
import docspell.backend.fulltext.CreateIndex
|
||||||
import docspell.common._
|
import docspell.common._
|
||||||
import docspell.ftsclient.FtsClient
|
import docspell.ftsclient.FtsClient
|
||||||
import docspell.store.queries.{QAttachment, QItem, QMoveAttachment}
|
import docspell.store.queries.{QAttachment, QItem, QMoveAttachment}
|
||||||
@ -212,6 +213,7 @@ object OItem {
|
|||||||
def apply[F[_]: Async](
|
def apply[F[_]: Async](
|
||||||
store: Store[F],
|
store: Store[F],
|
||||||
fts: FtsClient[F],
|
fts: FtsClient[F],
|
||||||
|
createIndex: CreateIndex[F],
|
||||||
queue: JobQueue[F],
|
queue: JobQueue[F],
|
||||||
joex: OJoex[F]
|
joex: OJoex[F]
|
||||||
): Resource[F, OItem[F]] =
|
): Resource[F, OItem[F]] =
|
||||||
@ -588,12 +590,13 @@ object OItem {
|
|||||||
items: NonEmptyList[Ident],
|
items: NonEmptyList[Ident],
|
||||||
collective: Ident
|
collective: Ident
|
||||||
): F[UpdateResult] =
|
): F[UpdateResult] =
|
||||||
UpdateResult.fromUpdate(
|
UpdateResult.fromUpdate(for {
|
||||||
store
|
n <- store
|
||||||
.transact(
|
.transact(
|
||||||
RItem.restoreStateForCollective(items, ItemState.Created, collective)
|
RItem.restoreStateForCollective(items, ItemState.Created, collective)
|
||||||
)
|
)
|
||||||
)
|
_ <- createIndex.reIndexData(logger, collective.some, items.some, 10)
|
||||||
|
} yield n)
|
||||||
|
|
||||||
def setItemDate(
|
def setItemDate(
|
||||||
items: NonEmptyList[Ident],
|
items: NonEmptyList[Ident],
|
||||||
@ -628,7 +631,10 @@ object OItem {
|
|||||||
} yield n
|
} yield n
|
||||||
|
|
||||||
def setDeletedState(items: NonEmptyList[Ident], collective: Ident): F[Int] =
|
def setDeletedState(items: NonEmptyList[Ident], collective: Ident): F[Int] =
|
||||||
store.transact(RItem.setState(items, collective, ItemState.Deleted))
|
for {
|
||||||
|
n <- store.transact(RItem.setState(items, collective, ItemState.Deleted))
|
||||||
|
_ <- items.traverse(id => fts.removeItem(logger, id))
|
||||||
|
} yield n
|
||||||
|
|
||||||
def getProposals(item: Ident, collective: Ident): F[MetaProposalList] =
|
def getProposals(item: Ident, collective: Ident): F[MetaProposalList] =
|
||||||
store.transact(QAttachment.getMetaProposals(item, collective))
|
store.transact(QAttachment.getMetaProposals(item, collective))
|
||||||
|
@ -231,13 +231,16 @@ object OSimpleSearch {
|
|||||||
case Some(ftq) if settings.useFTS =>
|
case Some(ftq) if settings.useFTS =>
|
||||||
if (q.isEmpty) {
|
if (q.isEmpty) {
|
||||||
logger.debug(s"Using index only search: $fulltextQuery")
|
logger.debug(s"Using index only search: $fulltextQuery")
|
||||||
fts
|
if (settings.searchMode == SearchMode.Trashed)
|
||||||
.findIndexOnly(settings.maxNoteLen)(
|
Items.ftsItemsFull(true)(Vector.empty).pure[F]
|
||||||
OFulltext.FtsInput(ftq),
|
else
|
||||||
q.fix.account,
|
fts
|
||||||
settings.batch
|
.findIndexOnly(settings.maxNoteLen)(
|
||||||
)
|
OFulltext.FtsInput(ftq),
|
||||||
.map(Items.ftsItemsFull(true))
|
q.fix.account,
|
||||||
|
settings.batch
|
||||||
|
)
|
||||||
|
.map(Items.ftsItemsFull(true))
|
||||||
} else if (settings.resolveDetails) {
|
} else if (settings.resolveDetails) {
|
||||||
logger.debug(
|
logger.debug(
|
||||||
s"Using index+sql search with tags: $validItemQuery / $fulltextQuery"
|
s"Using index+sql search with tags: $validItemQuery / $fulltextQuery"
|
||||||
|
@ -72,4 +72,5 @@ object TextData {
|
|||||||
notes: Option[String]
|
notes: Option[String]
|
||||||
): TextData =
|
): TextData =
|
||||||
Item(item, collective, folder, name, notes)
|
Item(item, collective, folder, name, notes)
|
||||||
|
|
||||||
}
|
}
|
||||||
|
@ -13,6 +13,7 @@ import cats.implicits._
|
|||||||
import fs2.concurrent.SignallingRef
|
import fs2.concurrent.SignallingRef
|
||||||
|
|
||||||
import docspell.analysis.TextAnalyser
|
import docspell.analysis.TextAnalyser
|
||||||
|
import docspell.backend.fulltext.CreateIndex
|
||||||
import docspell.backend.ops._
|
import docspell.backend.ops._
|
||||||
import docspell.common._
|
import docspell.common._
|
||||||
import docspell.ftsclient.FtsClient
|
import docspell.ftsclient.FtsClient
|
||||||
@ -116,7 +117,8 @@ object JoexAppImpl {
|
|||||||
joex <- OJoex(client, store)
|
joex <- OJoex(client, store)
|
||||||
upload <- OUpload(store, queue, cfg.files, joex)
|
upload <- OUpload(store, queue, cfg.files, joex)
|
||||||
fts <- createFtsClient(cfg)(httpClient)
|
fts <- createFtsClient(cfg)(httpClient)
|
||||||
itemOps <- OItem(store, fts, queue, joex)
|
createIndex <- CreateIndex.resource(fts, store)
|
||||||
|
itemOps <- OItem(store, fts, createIndex, queue, joex)
|
||||||
itemSearchOps <- OItemSearch(store)
|
itemSearchOps <- OItemSearch(store)
|
||||||
analyser <- TextAnalyser.create[F](cfg.textAnalysis.textAnalysisConfig)
|
analyser <- TextAnalyser.create[F](cfg.textAnalysis.textAnalysisConfig)
|
||||||
regexNer <- RegexNerFile(cfg.textAnalysis.regexNerFileConfig, store)
|
regexNer <- RegexNerFile(cfg.textAnalysis.regexNerFileConfig, store)
|
||||||
@ -155,14 +157,14 @@ object JoexAppImpl {
|
|||||||
.withTask(
|
.withTask(
|
||||||
JobTask.json(
|
JobTask.json(
|
||||||
MigrationTask.taskName,
|
MigrationTask.taskName,
|
||||||
MigrationTask[F](cfg.fullTextSearch, fts),
|
MigrationTask[F](cfg.fullTextSearch, fts, createIndex),
|
||||||
MigrationTask.onCancel[F]
|
MigrationTask.onCancel[F]
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
.withTask(
|
.withTask(
|
||||||
JobTask.json(
|
JobTask.json(
|
||||||
ReIndexTask.taskName,
|
ReIndexTask.taskName,
|
||||||
ReIndexTask[F](cfg.fullTextSearch, fts),
|
ReIndexTask[F](cfg.fullTextSearch, fts, createIndex),
|
||||||
ReIndexTask.onCancel[F]
|
ReIndexTask.onCancel[F]
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
|
@ -6,6 +6,7 @@
|
|||||||
|
|
||||||
package docspell.joex.fts
|
package docspell.joex.fts
|
||||||
|
|
||||||
|
import docspell.backend.fulltext.CreateIndex
|
||||||
import docspell.common.Logger
|
import docspell.common.Logger
|
||||||
import docspell.ftsclient.FtsClient
|
import docspell.ftsclient.FtsClient
|
||||||
import docspell.joex.Config
|
import docspell.joex.Config
|
||||||
@ -15,6 +16,7 @@ import docspell.store.Store
|
|||||||
case class FtsContext[F[_]](
|
case class FtsContext[F[_]](
|
||||||
cfg: Config.FullTextSearch,
|
cfg: Config.FullTextSearch,
|
||||||
store: Store[F],
|
store: Store[F],
|
||||||
|
fulltext: CreateIndex[F],
|
||||||
fts: FtsClient[F],
|
fts: FtsClient[F],
|
||||||
logger: Logger[F]
|
logger: Logger[F]
|
||||||
)
|
)
|
||||||
@ -24,7 +26,8 @@ object FtsContext {
|
|||||||
def apply[F[_]](
|
def apply[F[_]](
|
||||||
cfg: Config.FullTextSearch,
|
cfg: Config.FullTextSearch,
|
||||||
fts: FtsClient[F],
|
fts: FtsClient[F],
|
||||||
|
fulltext: CreateIndex[F],
|
||||||
ctx: Context[F, _]
|
ctx: Context[F, _]
|
||||||
): FtsContext[F] =
|
): FtsContext[F] =
|
||||||
FtsContext(cfg, ctx.store, fts, ctx.logger)
|
FtsContext(cfg, ctx.store, fulltext, fts, ctx.logger)
|
||||||
}
|
}
|
||||||
|
@ -10,11 +10,11 @@ import cats._
|
|||||||
import cats.data.{Kleisli, NonEmptyList}
|
import cats.data.{Kleisli, NonEmptyList}
|
||||||
import cats.implicits._
|
import cats.implicits._
|
||||||
|
|
||||||
|
import docspell.backend.fulltext.CreateIndex
|
||||||
import docspell.common._
|
import docspell.common._
|
||||||
import docspell.ftsclient._
|
import docspell.ftsclient._
|
||||||
import docspell.joex.Config
|
import docspell.joex.Config
|
||||||
import docspell.joex.scheduler.Context
|
import docspell.joex.scheduler.Context
|
||||||
import docspell.store.queries.{QAttachment, QItem}
|
|
||||||
|
|
||||||
object FtsWork {
|
object FtsWork {
|
||||||
import syntax._
|
import syntax._
|
||||||
@ -88,36 +88,8 @@ object FtsWork {
|
|||||||
log[F](_.info("Inserting all data to index")) ++ FtsWork
|
log[F](_.info("Inserting all data to index")) ++ FtsWork
|
||||||
.all(
|
.all(
|
||||||
FtsWork(ctx =>
|
FtsWork(ctx =>
|
||||||
ctx.fts.indexData(
|
ctx.fulltext
|
||||||
ctx.logger,
|
.reIndexData(ctx.logger, coll, None, ctx.cfg.migration.indexAllChunk)
|
||||||
ctx.store
|
|
||||||
.transact(
|
|
||||||
QAttachment
|
|
||||||
.allAttachmentMetaAndName(coll, ctx.cfg.migration.indexAllChunk)
|
|
||||||
)
|
|
||||||
.map(caa =>
|
|
||||||
TextData
|
|
||||||
.attachment(
|
|
||||||
caa.item,
|
|
||||||
caa.id,
|
|
||||||
caa.collective,
|
|
||||||
caa.folder,
|
|
||||||
caa.lang,
|
|
||||||
caa.name,
|
|
||||||
caa.content
|
|
||||||
)
|
|
||||||
)
|
|
||||||
)
|
|
||||||
),
|
|
||||||
FtsWork(ctx =>
|
|
||||||
ctx.fts.indexData(
|
|
||||||
ctx.logger,
|
|
||||||
ctx.store
|
|
||||||
.transact(QItem.allNameAndNotes(coll, ctx.cfg.migration.indexAllChunk * 5))
|
|
||||||
.map(nn =>
|
|
||||||
TextData.item(nn.id, nn.collective, nn.folder, Option(nn.name), nn.notes)
|
|
||||||
)
|
|
||||||
)
|
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
|
|
||||||
@ -133,9 +105,10 @@ object FtsWork {
|
|||||||
|
|
||||||
def forContext(
|
def forContext(
|
||||||
cfg: Config.FullTextSearch,
|
cfg: Config.FullTextSearch,
|
||||||
fts: FtsClient[F]
|
fts: FtsClient[F],
|
||||||
|
fulltext: CreateIndex[F]
|
||||||
): Kleisli[F, Context[F, _], Unit] =
|
): Kleisli[F, Context[F, _], Unit] =
|
||||||
mt.local(ctx => FtsContext(cfg, fts, ctx))
|
mt.local(ctx => FtsContext(cfg, fts, fulltext, ctx))
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -11,6 +11,7 @@ import cats.effect._
|
|||||||
import cats.implicits._
|
import cats.implicits._
|
||||||
import cats.{Applicative, FlatMap, Traverse}
|
import cats.{Applicative, FlatMap, Traverse}
|
||||||
|
|
||||||
|
import docspell.backend.fulltext.CreateIndex
|
||||||
import docspell.common._
|
import docspell.common._
|
||||||
import docspell.ftsclient._
|
import docspell.ftsclient._
|
||||||
import docspell.joex.Config
|
import docspell.joex.Config
|
||||||
@ -38,9 +39,10 @@ object Migration {
|
|||||||
cfg: Config.FullTextSearch,
|
cfg: Config.FullTextSearch,
|
||||||
fts: FtsClient[F],
|
fts: FtsClient[F],
|
||||||
store: Store[F],
|
store: Store[F],
|
||||||
|
createIndex: CreateIndex[F],
|
||||||
logger: Logger[F]
|
logger: Logger[F]
|
||||||
): Kleisli[F, List[Migration[F]], Unit] = {
|
): Kleisli[F, List[Migration[F]], Unit] = {
|
||||||
val ctx = FtsContext(cfg, store, fts, logger)
|
val ctx = FtsContext(cfg, store, createIndex, fts, logger)
|
||||||
Kleisli { migs =>
|
Kleisli { migs =>
|
||||||
if (migs.isEmpty) logger.info("No fulltext search migrations to run.")
|
if (migs.isEmpty) logger.info("No fulltext search migrations to run.")
|
||||||
else Traverse[List].sequence(migs.map(applySingle[F](ctx))).map(_ => ())
|
else Traverse[List].sequence(migs.map(applySingle[F](ctx))).map(_ => ())
|
||||||
|
@ -9,6 +9,7 @@ package docspell.joex.fts
|
|||||||
import cats.effect._
|
import cats.effect._
|
||||||
import cats.implicits._
|
import cats.implicits._
|
||||||
|
|
||||||
|
import docspell.backend.fulltext.CreateIndex
|
||||||
import docspell.common._
|
import docspell.common._
|
||||||
import docspell.ftsclient._
|
import docspell.ftsclient._
|
||||||
import docspell.joex.Config
|
import docspell.joex.Config
|
||||||
@ -20,7 +21,8 @@ object MigrationTask {
|
|||||||
|
|
||||||
def apply[F[_]: Async](
|
def apply[F[_]: Async](
|
||||||
cfg: Config.FullTextSearch,
|
cfg: Config.FullTextSearch,
|
||||||
fts: FtsClient[F]
|
fts: FtsClient[F],
|
||||||
|
createIndex: CreateIndex[F]
|
||||||
): Task[F, Unit, Unit] =
|
): Task[F, Unit, Unit] =
|
||||||
Task
|
Task
|
||||||
.log[F, Unit](_.info(s"Running full-text-index migrations now"))
|
.log[F, Unit](_.info(s"Running full-text-index migrations now"))
|
||||||
@ -28,7 +30,7 @@ object MigrationTask {
|
|||||||
Task(ctx =>
|
Task(ctx =>
|
||||||
for {
|
for {
|
||||||
migs <- migrationTasks[F](fts)
|
migs <- migrationTasks[F](fts)
|
||||||
res <- Migration[F](cfg, fts, ctx.store, ctx.logger).run(migs)
|
res <- Migration[F](cfg, fts, ctx.store, createIndex, ctx.logger).run(migs)
|
||||||
} yield res
|
} yield res
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
|
@ -8,6 +8,7 @@ package docspell.joex.fts
|
|||||||
|
|
||||||
import cats.effect._
|
import cats.effect._
|
||||||
|
|
||||||
|
import docspell.backend.fulltext.CreateIndex
|
||||||
import docspell.common._
|
import docspell.common._
|
||||||
import docspell.ftsclient._
|
import docspell.ftsclient._
|
||||||
import docspell.joex.Config
|
import docspell.joex.Config
|
||||||
@ -22,12 +23,15 @@ object ReIndexTask {
|
|||||||
|
|
||||||
def apply[F[_]: Async](
|
def apply[F[_]: Async](
|
||||||
cfg: Config.FullTextSearch,
|
cfg: Config.FullTextSearch,
|
||||||
fts: FtsClient[F]
|
fts: FtsClient[F],
|
||||||
|
fulltext: CreateIndex[F]
|
||||||
): Task[F, Args, Unit] =
|
): Task[F, Args, Unit] =
|
||||||
Task
|
Task
|
||||||
.log[F, Args](_.info(s"Running full-text re-index now"))
|
.log[F, Args](_.info(s"Running full-text re-index now"))
|
||||||
.flatMap(_ =>
|
.flatMap(_ =>
|
||||||
Task(ctx => clearData[F](ctx.args.collective).forContext(cfg, fts).run(ctx))
|
Task(ctx =>
|
||||||
|
clearData[F](ctx.args.collective).forContext(cfg, fts, fulltext).run(ctx)
|
||||||
|
)
|
||||||
)
|
)
|
||||||
|
|
||||||
def onCancel[F[_]]: Task[F, Args, Unit] =
|
def onCancel[F[_]]: Task[F, Args, Unit] =
|
||||||
|
@ -7,6 +7,7 @@
|
|||||||
package docspell.store.queries
|
package docspell.store.queries
|
||||||
|
|
||||||
import cats.data.OptionT
|
import cats.data.OptionT
|
||||||
|
import cats.data.{NonEmptyList => Nel}
|
||||||
import cats.effect.Sync
|
import cats.effect.Sync
|
||||||
import cats.implicits._
|
import cats.implicits._
|
||||||
import fs2.Stream
|
import fs2.Stream
|
||||||
@ -174,6 +175,7 @@ object QAttachment {
|
|||||||
)
|
)
|
||||||
def allAttachmentMetaAndName(
|
def allAttachmentMetaAndName(
|
||||||
coll: Option[Ident],
|
coll: Option[Ident],
|
||||||
|
itemIds: Option[Nel[Ident]],
|
||||||
chunkSize: Int
|
chunkSize: Int
|
||||||
): Stream[ConnectionIO, ContentAndName] =
|
): Stream[ConnectionIO, ContentAndName] =
|
||||||
Select(
|
Select(
|
||||||
@ -190,8 +192,11 @@ object QAttachment {
|
|||||||
.innerJoin(am, am.id === a.id)
|
.innerJoin(am, am.id === a.id)
|
||||||
.innerJoin(item, item.id === a.itemId)
|
.innerJoin(item, item.id === a.itemId)
|
||||||
.innerJoin(c, c.id === item.cid)
|
.innerJoin(c, c.id === item.cid)
|
||||||
).where(coll.map(cid => item.cid === cid))
|
).where(
|
||||||
.build
|
item.state.in(ItemState.validStates) &&?
|
||||||
|
itemIds.map(ids => item.id.in(ids)) &&?
|
||||||
|
coll.map(cid => item.cid === cid)
|
||||||
|
).build
|
||||||
.query[ContentAndName]
|
.query[ContentAndName]
|
||||||
.streamWithChunkSize(chunkSize)
|
.streamWithChunkSize(chunkSize)
|
||||||
|
|
||||||
|
@ -502,6 +502,7 @@ object QItem {
|
|||||||
.leftJoin(m3, m3.id === r.fileId),
|
.leftJoin(m3, m3.id === r.fileId),
|
||||||
where(
|
where(
|
||||||
i.cid === collective &&
|
i.cid === collective &&
|
||||||
|
i.state.in(ItemState.validStates) &&
|
||||||
Condition.Or(fms.map(m => m.checksum === checksum)) &&?
|
Condition.Or(fms.map(m => m.checksum === checksum)) &&?
|
||||||
Nel
|
Nel
|
||||||
.fromList(excludeFileMeta.toList)
|
.fromList(excludeFileMeta.toList)
|
||||||
@ -519,6 +520,7 @@ object QItem {
|
|||||||
)
|
)
|
||||||
def allNameAndNotes(
|
def allNameAndNotes(
|
||||||
coll: Option[Ident],
|
coll: Option[Ident],
|
||||||
|
itemIds: Option[Nel[Ident]],
|
||||||
chunkSize: Int
|
chunkSize: Int
|
||||||
): Stream[ConnectionIO, NameAndNotes] = {
|
): Stream[ConnectionIO, NameAndNotes] = {
|
||||||
val i = RItem.as("i")
|
val i = RItem.as("i")
|
||||||
@ -526,8 +528,11 @@ object QItem {
|
|||||||
Select(
|
Select(
|
||||||
select(i.id, i.cid, i.folder, i.name, i.notes),
|
select(i.id, i.cid, i.folder, i.name, i.notes),
|
||||||
from(i)
|
from(i)
|
||||||
).where(coll.map(cid => i.cid === cid))
|
).where(
|
||||||
.build
|
i.state.in(ItemState.validStates) &&?
|
||||||
|
itemIds.map(ids => i.id.in(ids)) &&?
|
||||||
|
coll.map(cid => i.cid === cid)
|
||||||
|
).build
|
||||||
.query[NameAndNotes]
|
.query[NameAndNotes]
|
||||||
.streamWithChunkSize(chunkSize)
|
.streamWithChunkSize(chunkSize)
|
||||||
}
|
}
|
||||||
|
Loading…
x
Reference in New Issue
Block a user