|
|
@ -1,8 +1,6 @@ |
|
|
|
package io.legado.app.model.localBook |
|
|
|
package io.legado.app.model.localBook |
|
|
|
|
|
|
|
|
|
|
|
import android.net.Uri |
|
|
|
import android.net.Uri |
|
|
|
import android.system.Os |
|
|
|
|
|
|
|
import android.system.OsConstants.SEEK_SET |
|
|
|
|
|
|
|
import io.legado.app.data.appDb |
|
|
|
import io.legado.app.data.appDb |
|
|
|
import io.legado.app.data.entities.Book |
|
|
|
import io.legado.app.data.entities.Book |
|
|
|
import io.legado.app.data.entities.BookChapter |
|
|
|
import io.legado.app.data.entities.BookChapter |
|
|
@ -11,8 +9,10 @@ import io.legado.app.help.DefaultData |
|
|
|
import io.legado.app.model.localBook.LocalBook.cacheFolder |
|
|
|
import io.legado.app.model.localBook.LocalBook.cacheFolder |
|
|
|
import io.legado.app.utils.* |
|
|
|
import io.legado.app.utils.* |
|
|
|
import splitties.init.appCtx |
|
|
|
import splitties.init.appCtx |
|
|
|
import java.io.* |
|
|
|
import java.io.File |
|
|
|
import java.nio.ByteBuffer |
|
|
|
import java.io.FileInputStream |
|
|
|
|
|
|
|
import java.io.FileNotFoundException |
|
|
|
|
|
|
|
import java.io.FileOutputStream |
|
|
|
import java.nio.charset.Charset |
|
|
|
import java.nio.charset.Charset |
|
|
|
import java.util.regex.Matcher |
|
|
|
import java.util.regex.Matcher |
|
|
|
import java.util.regex.Pattern |
|
|
|
import java.util.regex.Pattern |
|
|
@ -24,73 +24,67 @@ class TextFile(private val book: Book) { |
|
|
|
|
|
|
|
|
|
|
|
@Throws(FileNotFoundException::class) |
|
|
|
@Throws(FileNotFoundException::class) |
|
|
|
fun getChapterList(): ArrayList<BookChapter> { |
|
|
|
fun getChapterList(): ArrayList<BookChapter> { |
|
|
|
return getBookFD(book).let { fd -> |
|
|
|
var rulePattern: Pattern? = null |
|
|
|
try { |
|
|
|
if (book.charset == null || book.tocUrl.isNotEmpty()) { |
|
|
|
val rulePattern = if (book.tocUrl.isNotEmpty()) { |
|
|
|
getBookInputStream(book).use { bis -> |
|
|
|
|
|
|
|
val buffer = ByteArray(BUFFER_SIZE) |
|
|
|
|
|
|
|
var blockContent: String |
|
|
|
|
|
|
|
bis.read(buffer) |
|
|
|
|
|
|
|
book.charset = EncodingDetect.getEncode(buffer) |
|
|
|
|
|
|
|
charset = book.fileCharset() |
|
|
|
|
|
|
|
blockContent = String(buffer, charset) |
|
|
|
|
|
|
|
rulePattern = if (book.tocUrl.isNotEmpty()) { |
|
|
|
Pattern.compile(book.tocUrl, Pattern.MULTILINE) |
|
|
|
Pattern.compile(book.tocUrl, Pattern.MULTILINE) |
|
|
|
} else { |
|
|
|
} else { |
|
|
|
tocRules.addAll(getTocRules()) |
|
|
|
tocRules.addAll(getTocRules()) |
|
|
|
null |
|
|
|
if (blockContent.isEmpty()) { |
|
|
|
|
|
|
|
bis.read(buffer) |
|
|
|
|
|
|
|
book.charset = EncodingDetect.getEncode(buffer) |
|
|
|
|
|
|
|
blockContent = String(buffer, charset) |
|
|
|
|
|
|
|
} |
|
|
|
|
|
|
|
getTocRule(blockContent)?.let { |
|
|
|
|
|
|
|
Pattern.compile(it.rule, Pattern.MULTILINE) |
|
|
|
} |
|
|
|
} |
|
|
|
analyze(fd, book, rulePattern) |
|
|
|
|
|
|
|
} finally { |
|
|
|
|
|
|
|
if (fd.valid()) { |
|
|
|
|
|
|
|
Os.close(fd) |
|
|
|
|
|
|
|
} |
|
|
|
} |
|
|
|
} |
|
|
|
} |
|
|
|
} |
|
|
|
} |
|
|
|
|
|
|
|
return analyze(rulePattern) |
|
|
|
} |
|
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
private fun analyze( |
|
|
|
private fun analyze(pattern: Pattern?): ArrayList<BookChapter> { |
|
|
|
bookFd: FileDescriptor, |
|
|
|
|
|
|
|
book: Book, |
|
|
|
|
|
|
|
pattern: Pattern? |
|
|
|
|
|
|
|
): ArrayList<BookChapter> { |
|
|
|
|
|
|
|
val toc = arrayListOf<BookChapter>() |
|
|
|
val toc = arrayListOf<BookChapter>() |
|
|
|
|
|
|
|
getBookInputStream(book).use { bis -> |
|
|
|
var tocRule: TxtTocRule? = null |
|
|
|
var tocRule: TxtTocRule? = null |
|
|
|
val buffer = ByteArray(BUFFER_SIZE) |
|
|
|
val buffer = ByteArray(BUFFER_SIZE) |
|
|
|
var blockContent = "" |
|
|
|
var blockContent: String |
|
|
|
if (book.charset == null) { |
|
|
|
|
|
|
|
Os.lseek(bookFd, 0, SEEK_SET) |
|
|
|
|
|
|
|
val length = Os.read(bookFd, buffer, 0, BUFFER_SIZE) |
|
|
|
|
|
|
|
book.charset = EncodingDetect.getEncode(buffer) |
|
|
|
|
|
|
|
blockContent = String(buffer, 0, length, charset) |
|
|
|
|
|
|
|
charset = book.fileCharset() |
|
|
|
|
|
|
|
} |
|
|
|
|
|
|
|
val rulePattern = pattern ?: let { |
|
|
|
val rulePattern = pattern ?: let { |
|
|
|
if (blockContent.isEmpty()) { |
|
|
|
val length = bis.read(buffer) |
|
|
|
Os.lseek(bookFd, 0, SEEK_SET) |
|
|
|
bis.skip(-length.toLong()) |
|
|
|
val length = Os.read(bookFd, buffer, 0, BUFFER_SIZE) |
|
|
|
blockContent = String(buffer, charset) |
|
|
|
blockContent = String(buffer, 0, length, charset) |
|
|
|
|
|
|
|
} |
|
|
|
|
|
|
|
tocRule = getTocRule(blockContent) |
|
|
|
tocRule = getTocRule(blockContent) |
|
|
|
tocRule?.let { |
|
|
|
tocRule?.let { |
|
|
|
Pattern.compile(it.rule, Pattern.MULTILINE) |
|
|
|
Pattern.compile(it.rule, Pattern.MULTILINE) |
|
|
|
} |
|
|
|
} |
|
|
|
} |
|
|
|
} |
|
|
|
//加载章节 |
|
|
|
//加载章节 |
|
|
|
//获取到的块起始点,在文件中的位置 |
|
|
|
|
|
|
|
Os.lseek(bookFd, 0, SEEK_SET) |
|
|
|
|
|
|
|
var curOffset: Long = 0 |
|
|
|
var curOffset: Long = 0 |
|
|
|
//block的个数 |
|
|
|
//block的个数 |
|
|
|
var blockPos = 0 |
|
|
|
var blockPos = 0 |
|
|
|
//读取的长度 |
|
|
|
//读取的长度 |
|
|
|
var length: Int |
|
|
|
var length: Int |
|
|
|
var allLength = 0 |
|
|
|
|
|
|
|
val xx = ByteBuffer.allocate(BUFFER_SIZE) |
|
|
|
|
|
|
|
//获取文件中的数据到buffer,直到没有数据为止 |
|
|
|
//获取文件中的数据到buffer,直到没有数据为止 |
|
|
|
while (Os.read(bookFd, xx).also { length = it } > 0) { |
|
|
|
while (bis.read(buffer).also { length = it } > 0) { |
|
|
|
blockPos++ |
|
|
|
blockPos++ |
|
|
|
//如果存在Chapter |
|
|
|
//如果存在Chapter |
|
|
|
if (rulePattern != null) { |
|
|
|
if (rulePattern != null) { |
|
|
|
//将数据转换成String, 不能超过length |
|
|
|
//将数据转换成String, 不能超过length |
|
|
|
blockContent = String(xx.array(), 0, length, charset) |
|
|
|
blockContent = String(buffer, 0, length, charset) |
|
|
|
val lastN = blockContent.lastIndexOf("\n") |
|
|
|
val lastN = blockContent.lastIndexOf("\n") |
|
|
|
if (lastN > 0) { |
|
|
|
if (lastN > 0) { |
|
|
|
blockContent = blockContent.substring(0, lastN) |
|
|
|
blockContent = blockContent.substring(0, lastN) |
|
|
|
length = blockContent.toByteArray(charset).size |
|
|
|
val blockContentSize = blockContent.toByteArray(charset).size |
|
|
|
allLength += length |
|
|
|
bis.skip(-(length - blockContentSize).toLong()) |
|
|
|
Os.lseek(bookFd, allLength.toLong(), SEEK_SET) |
|
|
|
length = blockContentSize |
|
|
|
} |
|
|
|
} |
|
|
|
//当前Block下使过的String的指针 |
|
|
|
//当前Block下使过的String的指针 |
|
|
|
var seekPos = 0 |
|
|
|
var seekPos = 0 |
|
|
@ -106,7 +100,8 @@ class TextFile(private val book: Book) { |
|
|
|
if (curOffset + chapterLength - lastStart > 50000 && pattern == null) { |
|
|
|
if (curOffset + chapterLength - lastStart > 50000 && pattern == null) { |
|
|
|
//移除不匹配的规则 |
|
|
|
//移除不匹配的规则 |
|
|
|
tocRules.remove(tocRule) |
|
|
|
tocRules.remove(tocRule) |
|
|
|
return analyze(bookFd, book, null) |
|
|
|
bis.close() |
|
|
|
|
|
|
|
return analyze(null) |
|
|
|
} |
|
|
|
} |
|
|
|
//如果 seekPos == 0 && nextChapterPos != 0 表示当前block处前面有一段内容 |
|
|
|
//如果 seekPos == 0 && nextChapterPos != 0 表示当前block处前面有一段内容 |
|
|
|
//第一种情况一定是序章 第二种情况是上一个章节的内容 |
|
|
|
//第一种情况一定是序章 第二种情况是上一个章节的内容 |
|
|
@ -162,7 +157,8 @@ class TextFile(private val book: Book) { |
|
|
|
if (seekPos == 0 && length > 50000 && pattern == null) { |
|
|
|
if (seekPos == 0 && length > 50000 && pattern == null) { |
|
|
|
//移除不匹配的规则 |
|
|
|
//移除不匹配的规则 |
|
|
|
tocRules.remove(tocRule) |
|
|
|
tocRules.remove(tocRule) |
|
|
|
return analyze(bookFd, book, null) |
|
|
|
bis.close() |
|
|
|
|
|
|
|
return analyze(null) |
|
|
|
} |
|
|
|
} |
|
|
|
} else { //进行本地虚拟分章 |
|
|
|
} else { //进行本地虚拟分章 |
|
|
|
//章节在buffer的偏移量 |
|
|
|
//章节在buffer的偏移量 |
|
|
@ -218,6 +214,10 @@ class TextFile(private val book: Book) { |
|
|
|
System.runFinalization() |
|
|
|
System.runFinalization() |
|
|
|
} |
|
|
|
} |
|
|
|
} |
|
|
|
} |
|
|
|
|
|
|
|
tocRule?.let { |
|
|
|
|
|
|
|
book.tocUrl = it.rule |
|
|
|
|
|
|
|
} |
|
|
|
|
|
|
|
} |
|
|
|
for (i in toc.indices) { |
|
|
|
for (i in toc.indices) { |
|
|
|
val bean = toc[i] |
|
|
|
val bean = toc[i] |
|
|
|
bean.index = i |
|
|
|
bean.index = i |
|
|
@ -229,9 +229,7 @@ class TextFile(private val book: Book) { |
|
|
|
|
|
|
|
|
|
|
|
System.gc() |
|
|
|
System.gc() |
|
|
|
System.runFinalization() |
|
|
|
System.runFinalization() |
|
|
|
tocRule?.let { |
|
|
|
|
|
|
|
book.tocUrl = it.rule |
|
|
|
|
|
|
|
} |
|
|
|
|
|
|
|
return toc |
|
|
|
return toc |
|
|
|
} |
|
|
|
} |
|
|
|
|
|
|
|
|
|
|
@ -241,18 +239,22 @@ class TextFile(private val book: Book) { |
|
|
|
private fun getTocRule(content: String): TxtTocRule? { |
|
|
|
private fun getTocRule(content: String): TxtTocRule? { |
|
|
|
var txtTocRule: TxtTocRule? = null |
|
|
|
var txtTocRule: TxtTocRule? = null |
|
|
|
var maxCs = 0 |
|
|
|
var maxCs = 0 |
|
|
|
for (tocRule in tocRules) { |
|
|
|
val removeRules = hashSetOf<TxtTocRule>() |
|
|
|
|
|
|
|
tocRules.forEach { tocRule -> |
|
|
|
val pattern = Pattern.compile(tocRule.rule, Pattern.MULTILINE) |
|
|
|
val pattern = Pattern.compile(tocRule.rule, Pattern.MULTILINE) |
|
|
|
val matcher = pattern.matcher(content) |
|
|
|
val matcher = pattern.matcher(content) |
|
|
|
var cs = 0 |
|
|
|
var cs = 0 |
|
|
|
while (matcher.find()) { |
|
|
|
while (matcher.find()) { |
|
|
|
cs++ |
|
|
|
cs++ |
|
|
|
} |
|
|
|
} |
|
|
|
if (cs > maxCs) { |
|
|
|
if (cs == 0) { |
|
|
|
|
|
|
|
removeRules.add(tocRule) |
|
|
|
|
|
|
|
} else if (cs > maxCs) { |
|
|
|
maxCs = cs |
|
|
|
maxCs = cs |
|
|
|
txtTocRule = tocRule |
|
|
|
txtTocRule = tocRule |
|
|
|
} |
|
|
|
} |
|
|
|
} |
|
|
|
} |
|
|
|
|
|
|
|
tocRules.removeAll(removeRules) |
|
|
|
return txtTocRule |
|
|
|
return txtTocRule |
|
|
|
} |
|
|
|
} |
|
|
|
|
|
|
|
|
|
|
@ -274,22 +276,18 @@ class TextFile(private val book: Book) { |
|
|
|
@Throws(FileNotFoundException::class) |
|
|
|
@Throws(FileNotFoundException::class) |
|
|
|
fun getContent(book: Book, bookChapter: BookChapter): String { |
|
|
|
fun getContent(book: Book, bookChapter: BookChapter): String { |
|
|
|
val count = (bookChapter.end!! - bookChapter.start!!).toInt() |
|
|
|
val count = (bookChapter.end!! - bookChapter.start!!).toInt() |
|
|
|
val content = ByteArray(count) |
|
|
|
val buffer = ByteArray(count) |
|
|
|
getBookFD(book).let { fd -> |
|
|
|
getBookInputStream(book).use { bis -> |
|
|
|
try { |
|
|
|
bis.skip(bookChapter.start!!) |
|
|
|
Os.lseek(fd, bookChapter.start!!, SEEK_SET) |
|
|
|
bis.read(buffer) |
|
|
|
Os.read(fd, content, 0, count) |
|
|
|
|
|
|
|
} finally { |
|
|
|
|
|
|
|
Os.close(fd) |
|
|
|
|
|
|
|
} |
|
|
|
|
|
|
|
} |
|
|
|
} |
|
|
|
return String(content, book.fileCharset()) |
|
|
|
return String(buffer, book.fileCharset()) |
|
|
|
.substringAfter(bookChapter.title) |
|
|
|
.substringAfter(bookChapter.title) |
|
|
|
.replace("^[\\n\\s]+".toRegex(), " ") |
|
|
|
.replace("^[\\n\\s]+".toRegex(), " ") |
|
|
|
} |
|
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
@Throws(FileNotFoundException::class) |
|
|
|
@Throws(FileNotFoundException::class) |
|
|
|
private fun getBookFD(book: Book): FileDescriptor { |
|
|
|
private fun getBookInputStream(book: Book): FileInputStream { |
|
|
|
if (book.bookUrl.isContentScheme()) { |
|
|
|
if (book.bookUrl.isContentScheme()) { |
|
|
|
val uri = Uri.parse(book.bookUrl) |
|
|
|
val uri = Uri.parse(book.bookUrl) |
|
|
|
val bookFile = cacheFolder.getFile(book.name) |
|
|
|
val bookFile = cacheFolder.getFile(book.name) |
|
|
@ -302,10 +300,10 @@ class TextFile(private val book: Book) { |
|
|
|
} |
|
|
|
} |
|
|
|
} |
|
|
|
} |
|
|
|
} |
|
|
|
} |
|
|
|
return FileInputStream(bookFile).fd |
|
|
|
return FileInputStream(bookFile) |
|
|
|
//return appCtx.contentResolver.openFileDescriptor(uri, "r")!!.fileDescriptor |
|
|
|
//return appCtx.contentResolver.openFileDescriptor(uri, "r")!!.fileDescriptor |
|
|
|
} |
|
|
|
} |
|
|
|
return FileInputStream(File(book.bookUrl)).fd |
|
|
|
return FileInputStream(File(book.bookUrl)) |
|
|
|
} |
|
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
private fun getTocRules(): List<TxtTocRule> { |
|
|
|
private fun getTocRules(): List<TxtTocRule> { |
|
|
|