mirror of
https://github.com/TheAnachronism/docspell.git
synced 2025-04-05 19:09:32 +00:00
Merge pull request #857 from eikek/solr-migration
Improve Solr migration
This commit is contained in:
commit
ecc1e44e29
@ -17,12 +17,25 @@ import org.log4s.getLogger
|
|||||||
*/
|
*/
|
||||||
trait FtsClient[F[_]] {
|
trait FtsClient[F[_]] {
|
||||||
|
|
||||||
/** Initialization tasks. This is called exactly once at the very
|
/** Initialization tasks. This can be used to setup the fulltext
|
||||||
* beginning when initializing the full-text index and then never
|
* search engine. The implementation is expected to keep track of
|
||||||
* again (except when re-indexing everything). It may be used to
|
* run migrations, so that running these is idempotent. For
|
||||||
* setup the database.
|
* example, it may be run on each application start.
|
||||||
|
*
|
||||||
|
* Initialization may involve re-indexing all data, therefore it
|
||||||
|
* must run outside the scope of this client. The migration may
|
||||||
|
* include a task that applies any work and/or it can return a
|
||||||
|
* result indicating that after this task a re-index is necessary.
|
||||||
*/
|
*/
|
||||||
def initialize: List[FtsMigration[F]]
|
def initialize: F[List[FtsMigration[F]]]
|
||||||
|
|
||||||
|
/** A list of initialization tasks that can be run when re-creating
|
||||||
|
* the index.
|
||||||
|
*
|
||||||
|
* This is not run on startup, but only when required, for example
|
||||||
|
* when re-creating the entire index.
|
||||||
|
*/
|
||||||
|
def initializeNew: List[FtsMigration[F]]
|
||||||
|
|
||||||
/** Run a full-text search. */
|
/** Run a full-text search. */
|
||||||
def search(q: FtsQuery): F[FtsResult]
|
def search(q: FtsQuery): F[FtsResult]
|
||||||
@ -116,7 +129,10 @@ object FtsClient {
|
|||||||
new FtsClient[F] {
|
new FtsClient[F] {
|
||||||
private[this] val logger = Logger.log4s[F](getLogger)
|
private[this] val logger = Logger.log4s[F](getLogger)
|
||||||
|
|
||||||
def initialize: List[FtsMigration[F]] =
|
def initialize: F[List[FtsMigration[F]]] =
|
||||||
|
Sync[F].pure(Nil)
|
||||||
|
|
||||||
|
def initializeNew: List[FtsMigration[F]] =
|
||||||
Nil
|
Nil
|
||||||
|
|
||||||
def search(q: FtsQuery): F[FtsResult] =
|
def search(q: FtsQuery): F[FtsResult] =
|
||||||
|
@ -53,6 +53,37 @@ trait JsonCodec {
|
|||||||
): Encoder[TextData] =
|
): Encoder[TextData] =
|
||||||
Encoder(_.fold(ae.apply, ie.apply))
|
Encoder(_.fold(ae.apply, ie.apply))
|
||||||
|
|
||||||
|
implicit def versionDocEncoder: Encoder[VersionDoc] =
|
||||||
|
new Encoder[VersionDoc] {
|
||||||
|
final def apply(d: VersionDoc): Json =
|
||||||
|
Json.fromFields(
|
||||||
|
List(
|
||||||
|
(VersionDoc.Fields.id.name, d.id.asJson),
|
||||||
|
(
|
||||||
|
VersionDoc.Fields.currentVersion.name,
|
||||||
|
Map("set" -> d.currentVersion.asJson).asJson
|
||||||
|
)
|
||||||
|
)
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
implicit def decoderVersionDoc: Decoder[VersionDoc] =
|
||||||
|
new Decoder[VersionDoc] {
|
||||||
|
final def apply(c: HCursor): Decoder.Result[VersionDoc] =
|
||||||
|
for {
|
||||||
|
id <- c.get[String](VersionDoc.Fields.id.name)
|
||||||
|
version <- c.get[Int](VersionDoc.Fields.currentVersion.name)
|
||||||
|
} yield VersionDoc(id, version)
|
||||||
|
}
|
||||||
|
|
||||||
|
implicit def versionDocDecoder: Decoder[Option[VersionDoc]] =
|
||||||
|
new Decoder[Option[VersionDoc]] {
|
||||||
|
final def apply(c: HCursor): Decoder.Result[Option[VersionDoc]] =
|
||||||
|
c.downField("response")
|
||||||
|
.get[List[VersionDoc]]("docs")
|
||||||
|
.map(_.headOption)
|
||||||
|
}
|
||||||
|
|
||||||
implicit def docIdResultsDecoder: Decoder[DocIdResult] =
|
implicit def docIdResultsDecoder: Decoder[DocIdResult] =
|
||||||
new Decoder[DocIdResult] {
|
new Decoder[DocIdResult] {
|
||||||
final def apply(c: HCursor): Decoder.Result[DocIdResult] =
|
final def apply(c: HCursor): Decoder.Result[DocIdResult] =
|
||||||
|
@ -17,8 +17,11 @@ final class SolrFtsClient[F[_]: Effect](
|
|||||||
solrQuery: SolrQuery[F]
|
solrQuery: SolrQuery[F]
|
||||||
) extends FtsClient[F] {
|
) extends FtsClient[F] {
|
||||||
|
|
||||||
def initialize: List[FtsMigration[F]] =
|
def initialize: F[List[FtsMigration[F]]] =
|
||||||
solrSetup.setupSchema
|
solrSetup.remainingSetup.map(_.map(_.value))
|
||||||
|
|
||||||
|
def initializeNew: List[FtsMigration[F]] =
|
||||||
|
solrSetup.setupSchema.map(_.value)
|
||||||
|
|
||||||
def search(q: FtsQuery): F[FtsResult] =
|
def search(q: FtsQuery): F[FtsResult] =
|
||||||
solrQuery.query(q)
|
solrQuery.query(q)
|
||||||
|
@ -0,0 +1,73 @@
|
|||||||
|
package docspell.ftssolr
|
||||||
|
import cats.implicits._
|
||||||
|
import cats.{Applicative, Functor}
|
||||||
|
|
||||||
|
import docspell.common._
|
||||||
|
import docspell.ftsclient.FtsMigration
|
||||||
|
|
||||||
|
final case class SolrMigration[F[_]](value: FtsMigration[F], dataChangeOnly: Boolean) {
|
||||||
|
def isSchemaChange: Boolean = !dataChangeOnly
|
||||||
|
}
|
||||||
|
|
||||||
|
object SolrMigration {
|
||||||
|
private val solrEngine = Ident.unsafe("solr")
|
||||||
|
|
||||||
|
def deleteData[F[_]: Functor](
|
||||||
|
version: Int,
|
||||||
|
solrUpdate: SolrUpdate[F]
|
||||||
|
): SolrMigration[F] =
|
||||||
|
apply(version, "Delete all data", solrUpdate.delete("*:*", Option(0)))
|
||||||
|
|
||||||
|
def writeVersion[F[_]: Functor](
|
||||||
|
solrUpdate: SolrUpdate[F],
|
||||||
|
doc: VersionDoc
|
||||||
|
): SolrMigration[F] =
|
||||||
|
apply(
|
||||||
|
Int.MaxValue,
|
||||||
|
s"Write current version: ${doc.currentVersion}",
|
||||||
|
solrUpdate.updateVersionDoc(doc)
|
||||||
|
)
|
||||||
|
|
||||||
|
def reIndexAll[F[_]: Applicative](
|
||||||
|
versionNumber: Int,
|
||||||
|
description: String
|
||||||
|
): SolrMigration[F] =
|
||||||
|
SolrMigration(
|
||||||
|
FtsMigration(
|
||||||
|
versionNumber,
|
||||||
|
solrEngine,
|
||||||
|
description,
|
||||||
|
FtsMigration.Result.reIndexAll.pure[F]
|
||||||
|
),
|
||||||
|
true
|
||||||
|
)
|
||||||
|
|
||||||
|
def indexAll[F[_]: Applicative](
|
||||||
|
versionNumber: Int,
|
||||||
|
description: String
|
||||||
|
): SolrMigration[F] =
|
||||||
|
SolrMigration(
|
||||||
|
FtsMigration(
|
||||||
|
versionNumber,
|
||||||
|
solrEngine,
|
||||||
|
description,
|
||||||
|
FtsMigration.Result.indexAll.pure[F]
|
||||||
|
),
|
||||||
|
true
|
||||||
|
)
|
||||||
|
|
||||||
|
def apply[F[_]: Functor](
|
||||||
|
version: Int,
|
||||||
|
description: String,
|
||||||
|
task: F[Unit]
|
||||||
|
): SolrMigration[F] =
|
||||||
|
SolrMigration(
|
||||||
|
FtsMigration(
|
||||||
|
version,
|
||||||
|
solrEngine,
|
||||||
|
description,
|
||||||
|
task.map(_ => FtsMigration.Result.workDone)
|
||||||
|
),
|
||||||
|
false
|
||||||
|
)
|
||||||
|
}
|
@ -17,6 +17,8 @@ trait SolrQuery[F[_]] {
|
|||||||
def query(q: QueryData): F[FtsResult]
|
def query(q: QueryData): F[FtsResult]
|
||||||
|
|
||||||
def query(q: FtsQuery): F[FtsResult]
|
def query(q: FtsQuery): F[FtsResult]
|
||||||
|
|
||||||
|
def findVersionDoc(id: String): F[Option[VersionDoc]]
|
||||||
}
|
}
|
||||||
|
|
||||||
object SolrQuery {
|
object SolrQuery {
|
||||||
@ -54,6 +56,16 @@ object SolrQuery {
|
|||||||
)
|
)
|
||||||
query(fq)
|
query(fq)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
def findVersionDoc(id: String): F[Option[VersionDoc]] = {
|
||||||
|
val fields = List(
|
||||||
|
Field.id,
|
||||||
|
Field("current_version_i")
|
||||||
|
)
|
||||||
|
val query = QueryData(s"id:$id", "", 1, 0, fields, Map.empty)
|
||||||
|
val req = Method.POST(query.asJson, url)
|
||||||
|
client.expect[Option[VersionDoc]](req)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -4,7 +4,6 @@ import cats.effect._
|
|||||||
import cats.implicits._
|
import cats.implicits._
|
||||||
|
|
||||||
import docspell.common._
|
import docspell.common._
|
||||||
import docspell.ftsclient.FtsMigration
|
|
||||||
|
|
||||||
import _root_.io.circe._
|
import _root_.io.circe._
|
||||||
import _root_.io.circe.generic.semiauto._
|
import _root_.io.circe.generic.semiauto._
|
||||||
@ -16,12 +15,14 @@ import org.http4s.client.dsl.Http4sClientDsl
|
|||||||
|
|
||||||
trait SolrSetup[F[_]] {
|
trait SolrSetup[F[_]] {
|
||||||
|
|
||||||
def setupSchema: List[FtsMigration[F]]
|
def setupSchema: List[SolrMigration[F]]
|
||||||
|
|
||||||
|
def remainingSetup: F[List[SolrMigration[F]]]
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
object SolrSetup {
|
object SolrSetup {
|
||||||
private val solrEngine = Ident.unsafe("solr")
|
private val versionDocId = "6d8f09f4-8d7e-4bc9-98b8-7c89223b36dd"
|
||||||
|
|
||||||
def apply[F[_]: ConcurrentEffect](cfg: SolrConfig, client: Client[F]): SolrSetup[F] = {
|
def apply[F[_]: ConcurrentEffect](cfg: SolrConfig, client: Client[F]): SolrSetup[F] = {
|
||||||
val dsl = new Http4sClientDsl[F] {}
|
val dsl = new Http4sClientDsl[F] {}
|
||||||
@ -32,62 +33,76 @@ object SolrSetup {
|
|||||||
val url = (Uri.unsafeFromString(cfg.url.asString) / "schema")
|
val url = (Uri.unsafeFromString(cfg.url.asString) / "schema")
|
||||||
.withQueryParam("commitWithin", cfg.commitWithin.toString)
|
.withQueryParam("commitWithin", cfg.commitWithin.toString)
|
||||||
|
|
||||||
def setupSchema: List[FtsMigration[F]] =
|
def remainingSetup: F[List[SolrMigration[F]]] =
|
||||||
|
for {
|
||||||
|
current <- SolrQuery(cfg, client).findVersionDoc(versionDocId)
|
||||||
|
migs = current match {
|
||||||
|
case None => setupSchema
|
||||||
|
case Some(ver) =>
|
||||||
|
val verDoc =
|
||||||
|
VersionDoc(versionDocId, allMigrations.map(_.value.version).max)
|
||||||
|
val solrUp = SolrUpdate(cfg, client)
|
||||||
|
val remain = allMigrations.filter(v => v.value.version > ver.currentVersion)
|
||||||
|
if (remain.isEmpty) remain
|
||||||
|
else remain :+ SolrMigration.writeVersion(solrUp, verDoc)
|
||||||
|
}
|
||||||
|
} yield migs
|
||||||
|
|
||||||
|
def setupSchema: List[SolrMigration[F]] = {
|
||||||
|
val verDoc = VersionDoc(versionDocId, allMigrations.map(_.value.version).max)
|
||||||
|
val solrUp = SolrUpdate(cfg, client)
|
||||||
|
val writeVersion = SolrMigration.writeVersion(solrUp, verDoc)
|
||||||
|
val deleteAll = SolrMigration.deleteData(0, solrUp)
|
||||||
|
val indexAll = SolrMigration.indexAll[F](Int.MaxValue, "Index all data")
|
||||||
|
|
||||||
|
deleteAll :: (allMigrations
|
||||||
|
.filter(_.isSchemaChange) ::: List(indexAll, writeVersion))
|
||||||
|
}
|
||||||
|
|
||||||
|
private def allMigrations: List[SolrMigration[F]] =
|
||||||
List(
|
List(
|
||||||
FtsMigration[F](
|
SolrMigration[F](
|
||||||
1,
|
1,
|
||||||
solrEngine,
|
|
||||||
"Initialize",
|
"Initialize",
|
||||||
setupCoreSchema.map(_ => FtsMigration.Result.workDone)
|
setupCoreSchema
|
||||||
),
|
),
|
||||||
FtsMigration[F](
|
SolrMigration[F](
|
||||||
3,
|
2,
|
||||||
solrEngine,
|
|
||||||
"Add folder field",
|
"Add folder field",
|
||||||
addFolderField.map(_ => FtsMigration.Result.workDone)
|
addFolderField
|
||||||
),
|
),
|
||||||
FtsMigration[F](
|
SolrMigration.indexAll(3, "Index all from database after adding folder field"),
|
||||||
|
SolrMigration[F](
|
||||||
4,
|
4,
|
||||||
solrEngine,
|
|
||||||
"Index all from database",
|
|
||||||
FtsMigration.Result.indexAll.pure[F]
|
|
||||||
),
|
|
||||||
FtsMigration[F](
|
|
||||||
5,
|
|
||||||
solrEngine,
|
|
||||||
"Add content_fr field",
|
"Add content_fr field",
|
||||||
addContentField(Language.French).map(_ => FtsMigration.Result.workDone)
|
addContentField(Language.French)
|
||||||
),
|
),
|
||||||
FtsMigration[F](
|
SolrMigration
|
||||||
|
.indexAll(5, "Index all from database after adding french content field"),
|
||||||
|
SolrMigration[F](
|
||||||
6,
|
6,
|
||||||
solrEngine,
|
|
||||||
"Index all from database",
|
|
||||||
FtsMigration.Result.indexAll.pure[F]
|
|
||||||
),
|
|
||||||
FtsMigration[F](
|
|
||||||
7,
|
|
||||||
solrEngine,
|
|
||||||
"Add content_it field",
|
"Add content_it field",
|
||||||
addContentField(Language.Italian).map(_ => FtsMigration.Result.reIndexAll)
|
addContentField(Language.Italian)
|
||||||
),
|
),
|
||||||
FtsMigration[F](
|
SolrMigration.reIndexAll(7, "Re-Index after adding italian content field"),
|
||||||
|
SolrMigration[F](
|
||||||
8,
|
8,
|
||||||
solrEngine,
|
|
||||||
"Add content_es field",
|
"Add content_es field",
|
||||||
addContentField(Language.Spanish).map(_ => FtsMigration.Result.reIndexAll)
|
addContentField(Language.Spanish)
|
||||||
),
|
),
|
||||||
FtsMigration[F](
|
SolrMigration.reIndexAll(9, "Re-Index after adding spanish content field"),
|
||||||
9,
|
SolrMigration[F](
|
||||||
solrEngine,
|
|
||||||
"Add more content fields",
|
|
||||||
addMoreContentFields.map(_ => FtsMigration.Result.reIndexAll)
|
|
||||||
),
|
|
||||||
FtsMigration[F](
|
|
||||||
10,
|
10,
|
||||||
solrEngine,
|
"Add more content fields",
|
||||||
|
addMoreContentFields
|
||||||
|
),
|
||||||
|
SolrMigration.reIndexAll(11, "Re-Index after adding more content fields"),
|
||||||
|
SolrMigration[F](
|
||||||
|
12,
|
||||||
"Add latvian content field",
|
"Add latvian content field",
|
||||||
addContentField(Language.Latvian).map(_ => FtsMigration.Result.reIndexAll)
|
addContentField(Language.Latvian)
|
||||||
)
|
),
|
||||||
|
SolrMigration.reIndexAll(13, "Re-Index after adding latvian content field")
|
||||||
)
|
)
|
||||||
|
|
||||||
def addFolderField: F[Unit] =
|
def addFolderField: F[Unit] =
|
||||||
|
@ -23,6 +23,8 @@ trait SolrUpdate[F[_]] {
|
|||||||
|
|
||||||
def updateFolder(itemId: Ident, collective: Ident, folder: Option[Ident]): F[Unit]
|
def updateFolder(itemId: Ident, collective: Ident, folder: Option[Ident]): F[Unit]
|
||||||
|
|
||||||
|
def updateVersionDoc(doc: VersionDoc): F[Unit]
|
||||||
|
|
||||||
def delete(q: String, commitWithin: Option[Int]): F[Unit]
|
def delete(q: String, commitWithin: Option[Int]): F[Unit]
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -48,6 +50,11 @@ object SolrUpdate {
|
|||||||
client.expect[Unit](req)
|
client.expect[Unit](req)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
def updateVersionDoc(doc: VersionDoc): F[Unit] = {
|
||||||
|
val req = Method.POST(List(doc).asJson, url)
|
||||||
|
client.expect[Unit](req)
|
||||||
|
}
|
||||||
|
|
||||||
def updateFolder(
|
def updateFolder(
|
||||||
itemId: Ident,
|
itemId: Ident,
|
||||||
collective: Ident,
|
collective: Ident,
|
||||||
|
@ -0,0 +1,11 @@
|
|||||||
|
package docspell.ftssolr
|
||||||
|
|
||||||
|
final case class VersionDoc(id: String, currentVersion: Int)
|
||||||
|
|
||||||
|
object VersionDoc {
|
||||||
|
|
||||||
|
object Fields {
|
||||||
|
val id = Field("id")
|
||||||
|
val currentVersion = Field("current_version_i")
|
||||||
|
}
|
||||||
|
}
|
@ -11,20 +11,23 @@ import docspell.joex.scheduler.Context
|
|||||||
import docspell.store.queries.{QAttachment, QItem}
|
import docspell.store.queries.{QAttachment, QItem}
|
||||||
|
|
||||||
object FtsWork {
|
object FtsWork {
|
||||||
|
import syntax._
|
||||||
|
|
||||||
def apply[F[_]](f: FtsContext[F] => F[Unit]): FtsWork[F] =
|
def apply[F[_]](f: FtsContext[F] => F[Unit]): FtsWork[F] =
|
||||||
Kleisli(f)
|
Kleisli(f)
|
||||||
|
|
||||||
/** Runs all migration tasks unconditionally and inserts all data as last step. */
|
/** Runs migration tasks to re-create the index. */
|
||||||
def reInitializeTasks[F[_]: Monad]: FtsWork[F] =
|
def reInitializeTasks[F[_]: Monad]: FtsWork[F] =
|
||||||
FtsWork { ctx =>
|
FtsWork { ctx =>
|
||||||
val migrations =
|
val migrations = ctx.fts.initializeNew
|
||||||
ctx.fts.initialize.map(fm => fm.changeResult(_ => FtsMigration.Result.workDone))
|
|
||||||
|
|
||||||
NonEmptyList.fromList(migrations) match {
|
NonEmptyList.fromList(migrations) match {
|
||||||
case Some(nel) =>
|
case Some(nel) =>
|
||||||
nel
|
nel
|
||||||
.map(fm => from[F](fm.task))
|
.map(fm =>
|
||||||
.append(insertAll[F](None))
|
log[F](_.debug(s"Apply (${fm.engine.id}): ${fm.description}")) ++ from[F](
|
||||||
|
fm.task
|
||||||
|
)
|
||||||
|
)
|
||||||
.reduce(semigroup[F])
|
.reduce(semigroup[F])
|
||||||
.run(ctx)
|
.run(ctx)
|
||||||
case None =>
|
case None =>
|
||||||
@ -32,8 +35,6 @@ object FtsWork {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
|
||||||
*/
|
|
||||||
def from[F[_]: FlatMap: Applicative](t: F[FtsMigration.Result]): FtsWork[F] =
|
def from[F[_]: FlatMap: Applicative](t: F[FtsMigration.Result]): FtsWork[F] =
|
||||||
Kleisli.liftF(t).flatMap(transformResult[F])
|
Kleisli.liftF(t).flatMap(transformResult[F])
|
||||||
|
|
||||||
@ -65,16 +66,20 @@ object FtsWork {
|
|||||||
def log[F[_]](f: Logger[F] => F[Unit]): FtsWork[F] =
|
def log[F[_]](f: Logger[F] => F[Unit]): FtsWork[F] =
|
||||||
FtsWork(ctx => f(ctx.logger))
|
FtsWork(ctx => f(ctx.logger))
|
||||||
|
|
||||||
def clearIndex[F[_]](coll: Option[Ident]): FtsWork[F] =
|
def clearIndex[F[_]: FlatMap](coll: Option[Ident]): FtsWork[F] =
|
||||||
coll match {
|
coll match {
|
||||||
case Some(cid) =>
|
case Some(cid) =>
|
||||||
FtsWork(ctx => ctx.fts.clear(ctx.logger, cid))
|
log[F](_.debug(s"Clearing index data for collective '${cid.id}'")) ++ FtsWork(
|
||||||
|
ctx => ctx.fts.clear(ctx.logger, cid)
|
||||||
|
)
|
||||||
case None =>
|
case None =>
|
||||||
FtsWork(ctx => ctx.fts.clearAll(ctx.logger))
|
log[F](_.debug("Clearing all index data!")) ++ FtsWork(ctx =>
|
||||||
|
ctx.fts.clearAll(ctx.logger)
|
||||||
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
def insertAll[F[_]: FlatMap](coll: Option[Ident]): FtsWork[F] =
|
def insertAll[F[_]: FlatMap](coll: Option[Ident]): FtsWork[F] =
|
||||||
FtsWork
|
log[F](_.info("Inserting all data to index")) ++ FtsWork
|
||||||
.all(
|
.all(
|
||||||
FtsWork(ctx =>
|
FtsWork(ctx =>
|
||||||
ctx.fts.indexData(
|
ctx.fts.indexData(
|
||||||
|
@ -1,6 +1,6 @@
|
|||||||
package docspell.joex.fts
|
package docspell.joex.fts
|
||||||
|
|
||||||
import cats.data.{Kleisli, OptionT}
|
import cats.data.Kleisli
|
||||||
import cats.effect._
|
import cats.effect._
|
||||||
import cats.implicits._
|
import cats.implicits._
|
||||||
import cats.{Applicative, FlatMap, Traverse}
|
import cats.{Applicative, FlatMap, Traverse}
|
||||||
@ -8,13 +8,13 @@ import cats.{Applicative, FlatMap, Traverse}
|
|||||||
import docspell.common._
|
import docspell.common._
|
||||||
import docspell.ftsclient._
|
import docspell.ftsclient._
|
||||||
import docspell.joex.Config
|
import docspell.joex.Config
|
||||||
import docspell.store.records.RFtsMigration
|
import docspell.store.Store
|
||||||
import docspell.store.{AddResult, Store}
|
|
||||||
|
|
||||||
/** Migrating the index from the previous version to this version.
|
/** Migrating the index from the previous version to this version.
|
||||||
*
|
*
|
||||||
* The sql database stores the outcome of a migration task. If this
|
* The migration asks the fulltext search client for a list of
|
||||||
* task has already been applied, it is skipped.
|
* migration tasks to run. It may be empty when there is no migration
|
||||||
|
* required.
|
||||||
*/
|
*/
|
||||||
case class Migration[F[_]](
|
case class Migration[F[_]](
|
||||||
version: Int,
|
version: Int,
|
||||||
@ -35,41 +35,15 @@ object Migration {
|
|||||||
logger: Logger[F]
|
logger: Logger[F]
|
||||||
): Kleisli[F, List[Migration[F]], Unit] = {
|
): Kleisli[F, List[Migration[F]], Unit] = {
|
||||||
val ctx = FtsContext(cfg, store, fts, logger)
|
val ctx = FtsContext(cfg, store, fts, logger)
|
||||||
Kleisli(migs => Traverse[List].sequence(migs.map(applySingle[F](ctx))).map(_ => ()))
|
Kleisli { migs =>
|
||||||
|
if (migs.isEmpty) logger.info("No fulltext search migrations to run.")
|
||||||
|
else Traverse[List].sequence(migs.map(applySingle[F](ctx))).map(_ => ())
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
def applySingle[F[_]: Effect](ctx: FtsContext[F])(m: Migration[F]): F[Unit] = {
|
def applySingle[F[_]: Effect](ctx: FtsContext[F])(m: Migration[F]): F[Unit] =
|
||||||
val insertRecord: F[Option[RFtsMigration]] =
|
for {
|
||||||
for {
|
_ <- ctx.logger.info(s"Apply ${m.version}/${m.description}")
|
||||||
rec <- RFtsMigration.create(m.version, m.engine, m.description)
|
_ <- m.task.run(ctx)
|
||||||
res <- ctx.store.add(
|
} yield ()
|
||||||
RFtsMigration.insert(rec),
|
|
||||||
RFtsMigration.exists(m.version, m.engine)
|
|
||||||
)
|
|
||||||
ret <- res match {
|
|
||||||
case AddResult.Success => rec.some.pure[F]
|
|
||||||
case AddResult.EntityExists(_) => None.pure[F]
|
|
||||||
case AddResult.Failure(ex) => Effect[F].raiseError(ex)
|
|
||||||
}
|
|
||||||
} yield ret
|
|
||||||
|
|
||||||
(for {
|
|
||||||
_ <- OptionT.liftF(ctx.logger.info(s"Apply ${m.version}/${m.description}"))
|
|
||||||
rec <- OptionT(insertRecord)
|
|
||||||
res <- OptionT.liftF(m.task.run(ctx).attempt)
|
|
||||||
ret <- OptionT.liftF(res match {
|
|
||||||
case Right(()) => ().pure[F]
|
|
||||||
case Left(ex) =>
|
|
||||||
ctx.logger.error(ex)(
|
|
||||||
s"Applying index migration ${m.version}/${m.description} failed"
|
|
||||||
) *>
|
|
||||||
ctx.store.transact(RFtsMigration.deleteById(rec.id)) *> Effect[F]
|
|
||||||
.raiseError[Unit](
|
|
||||||
ex
|
|
||||||
)
|
|
||||||
})
|
|
||||||
} yield ret).getOrElseF(
|
|
||||||
ctx.logger.info(s"Migration ${m.version}/${m.description} already applied.")
|
|
||||||
)
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
@ -20,8 +20,10 @@ object MigrationTask {
|
|||||||
.log[F, Unit](_.info(s"Running full-text-index migrations now"))
|
.log[F, Unit](_.info(s"Running full-text-index migrations now"))
|
||||||
.flatMap(_ =>
|
.flatMap(_ =>
|
||||||
Task(ctx =>
|
Task(ctx =>
|
||||||
Migration[F](cfg, fts, ctx.store, ctx.logger)
|
for {
|
||||||
.run(migrationTasks[F](fts))
|
migs <- migrationTasks[F](fts)
|
||||||
|
res <- Migration[F](cfg, fts, ctx.store, ctx.logger).run(migs)
|
||||||
|
} yield res
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
|
|
||||||
@ -44,7 +46,7 @@ object MigrationTask {
|
|||||||
Some(DocspellSystem.migrationTaskTracker)
|
Some(DocspellSystem.migrationTaskTracker)
|
||||||
)
|
)
|
||||||
|
|
||||||
def migrationTasks[F[_]: Effect](fts: FtsClient[F]): List[Migration[F]] =
|
def migrationTasks[F[_]: Effect](fts: FtsClient[F]): F[List[Migration[F]]] =
|
||||||
fts.initialize.map(fm => Migration.from(fm))
|
fts.initialize.map(_.map(fm => Migration.from(fm)))
|
||||||
|
|
||||||
}
|
}
|
||||||
|
@ -40,12 +40,7 @@ object ReIndexTask {
|
|||||||
FtsWork.insertAll[F](collective)
|
FtsWork.insertAll[F](collective)
|
||||||
|
|
||||||
case None =>
|
case None =>
|
||||||
FtsWork
|
FtsWork.log[F](_.info("Running re-create index")) ++
|
||||||
.clearIndex(None)
|
|
||||||
.recoverWith(
|
|
||||||
FtsWork.log[F](_.info("Clearing data failed. Continue re-indexing."))
|
|
||||||
) ++
|
|
||||||
FtsWork.log[F](_.info("Running index initialize")) ++
|
|
||||||
FtsWork.reInitializeTasks[F]
|
FtsWork.reInitializeTasks[F]
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
@ -0,0 +1 @@
|
|||||||
|
DROP TABLE "fts_migration";
|
@ -0,0 +1 @@
|
|||||||
|
DROP TABLE `fts_migration`;
|
@ -0,0 +1 @@
|
|||||||
|
DROP TABLE "fts_migration";
|
@ -1,68 +0,0 @@
|
|||||||
package docspell.store.records
|
|
||||||
|
|
||||||
import cats.data.NonEmptyList
|
|
||||||
import cats.effect._
|
|
||||||
import cats.implicits._
|
|
||||||
|
|
||||||
import docspell.common._
|
|
||||||
import docspell.store.qb.DSL._
|
|
||||||
import docspell.store.qb._
|
|
||||||
|
|
||||||
import doobie._
|
|
||||||
import doobie.implicits._
|
|
||||||
|
|
||||||
final case class RFtsMigration(
|
|
||||||
id: Ident,
|
|
||||||
version: Int,
|
|
||||||
ftsEngine: Ident,
|
|
||||||
description: String,
|
|
||||||
created: Timestamp
|
|
||||||
)
|
|
||||||
|
|
||||||
object RFtsMigration {
|
|
||||||
|
|
||||||
def create[F[_]: Sync](
|
|
||||||
version: Int,
|
|
||||||
ftsEngine: Ident,
|
|
||||||
description: String
|
|
||||||
): F[RFtsMigration] =
|
|
||||||
for {
|
|
||||||
newId <- Ident.randomId[F]
|
|
||||||
now <- Timestamp.current[F]
|
|
||||||
} yield RFtsMigration(newId, version, ftsEngine, description, now)
|
|
||||||
|
|
||||||
final case class Table(alias: Option[String]) extends TableDef {
|
|
||||||
val tableName = "fts_migration"
|
|
||||||
|
|
||||||
val id = Column[Ident]("id", this)
|
|
||||||
val version = Column[Int]("version", this)
|
|
||||||
val ftsEngine = Column[Ident]("fts_engine", this)
|
|
||||||
val description = Column[String]("description", this)
|
|
||||||
val created = Column[Timestamp]("created", this)
|
|
||||||
|
|
||||||
val all = NonEmptyList.of[Column[_]](id, version, ftsEngine, description, created)
|
|
||||||
}
|
|
||||||
|
|
||||||
val T = Table(None)
|
|
||||||
def as(alias: String): Table =
|
|
||||||
Table(Some(alias))
|
|
||||||
|
|
||||||
def insert(v: RFtsMigration): ConnectionIO[Int] =
|
|
||||||
DML
|
|
||||||
.insertFragment(
|
|
||||||
T,
|
|
||||||
T.all,
|
|
||||||
Seq(fr"${v.id},${v.version},${v.ftsEngine},${v.description},${v.created}")
|
|
||||||
)
|
|
||||||
.updateWithLogHandler(LogHandler.nop)
|
|
||||||
.run
|
|
||||||
|
|
||||||
def exists(vers: Int, engine: Ident): ConnectionIO[Boolean] =
|
|
||||||
run(select(count(T.id)), from(T), T.version === vers && T.ftsEngine === engine)
|
|
||||||
.query[Int]
|
|
||||||
.unique
|
|
||||||
.map(_ > 0)
|
|
||||||
|
|
||||||
def deleteById(rId: Ident): ConnectionIO[Int] =
|
|
||||||
DML.delete(T, T.id === rId)
|
|
||||||
}
|
|
@ -100,7 +100,7 @@ itemsBar texts flags settings model =
|
|||||||
|
|
||||||
|
|
||||||
defaultMenuBar : Texts -> Flags -> UiSettings -> Model -> Html Msg
|
defaultMenuBar : Texts -> Flags -> UiSettings -> Model -> Html Msg
|
||||||
defaultMenuBar texts _ settings model =
|
defaultMenuBar texts flags settings model =
|
||||||
let
|
let
|
||||||
btnStyle =
|
btnStyle =
|
||||||
S.secondaryBasicButton ++ " text-sm"
|
S.secondaryBasicButton ++ " text-sm"
|
||||||
@ -127,12 +127,20 @@ defaultMenuBar texts _ settings model =
|
|||||||
, Maybe.map value searchInput
|
, Maybe.map value searchInput
|
||||||
|> Maybe.withDefault (value "")
|
|> Maybe.withDefault (value "")
|
||||||
, class (String.replace "rounded" "" S.textInput)
|
, class (String.replace "rounded" "" S.textInput)
|
||||||
, class "py-1 text-sm border-r-0 rounded-l"
|
, class "py-2 text-sm"
|
||||||
|
, if flags.config.fullTextSearchEnabled then
|
||||||
|
class " border-r-0 rounded-l"
|
||||||
|
|
||||||
|
else
|
||||||
|
class "border rounded"
|
||||||
]
|
]
|
||||||
[]
|
[]
|
||||||
, a
|
, a
|
||||||
[ class S.secondaryBasicButtonPlain
|
[ class S.secondaryBasicButtonPlain
|
||||||
, class "text-sm px-4 py-2 border rounded-r"
|
, class "text-sm px-4 py-2 border rounded-r"
|
||||||
|
, classList
|
||||||
|
[ ( "hidden", not flags.config.fullTextSearchEnabled )
|
||||||
|
]
|
||||||
, href "#"
|
, href "#"
|
||||||
, onClick ToggleSearchType
|
, onClick ToggleSearchType
|
||||||
]
|
]
|
||||||
|
@ -191,13 +191,17 @@ it is empty (the default), this call is disabled (all admin routes).
|
|||||||
Otherwise, the POST request will submit a system task that is executed
|
Otherwise, the POST request will submit a system task that is executed
|
||||||
by a joex instance eventually.
|
by a joex instance eventually.
|
||||||
|
|
||||||
Using this endpoint, the index will be re-created. This is sometimes
|
Using this endpoint, the entire index (including the schema) will be
|
||||||
necessary, for example if you upgrade SOLR or delete the core to
|
re-created. This is sometimes necessary, for example if you upgrade
|
||||||
provide a new one (see
|
SOLR or delete the core to provide a new one (see
|
||||||
[here](https://solr.apache.org/guide/8_4/reindexing.html) for
|
[here](https://solr.apache.org/guide/8_4/reindexing.html) for
|
||||||
details). Note that a collective can also re-index their data using a
|
details). Another way is to restart docspell (while clearing the
|
||||||
similiar endpoint; but this is only deleting their data and doesn't do
|
index). If docspell detects an empty index at startup, it will submit
|
||||||
a full re-index.
|
a task to build the index automatically.
|
||||||
|
|
||||||
|
Note that a collective can also re-index their data using a similiar
|
||||||
|
endpoint; but this is only deleting their data and doesn't do a full
|
||||||
|
re-index.
|
||||||
|
|
||||||
The solr index doesn't contain any new information, it can be
|
The solr index doesn't contain any new information, it can be
|
||||||
regenerated any time using the above REST call. Thus it doesn't need
|
regenerated any time using the above REST call. Thus it doesn't need
|
||||||
|
Loading…
x
Reference in New Issue
Block a user