Compare commits

..

6 Commits

Author SHA1 Message Date
cmumford
1a9648e1f5 Only compiling TrimSpace on linux.
Incorporated change by zmodem at https://github.com/google/leveldb/pull/310
to fix issue #310.

This change will only build TrimSace on linux to avoid unused function
warning/error.
-------------
Created by MOE: https://github.com/google/moe
MOE_MIGRATED_REVID=105323419
2015-10-13 14:06:08 -07:00
ndmatthews
f79f4180cc Let LevelDB use xcrun to determine Xcode.app path instead of using a hardcoded path.
This allows build agents to select from multiple Xcode installations.
-------------
Created by MOE: https://github.com/google/moe
MOE_MIGRATED_REVID=104859097
2015-10-13 14:05:43 -07:00
ssid
36fc955971 Add "approximate-memory-usage" property to leveldb::DB::GetProperty
The approximate RAM usage of the database is calculated from the memory
allocated for write buffers and the block cache. This is to give an
estimate of memory usage to leveldb clients.
-------------
Created by MOE: https://github.com/google/moe
MOE_MIGRATED_REVID=104222307
2015-10-13 14:05:30 -07:00
tzik
bb61e00815 Add leveldb::Cache::Prune
Prune() drops on-memory read cache of the database, so that the client can
relief its memory shortage.
-------------
Created by MOE: https://github.com/google/moe
MOE_MIGRATED_REVID=101335710
2015-10-13 14:05:21 -07:00
Chris Mumford
40c17c0b84 Will not reuse manifest if reuse_logs options is false.
Prior implementation would always try to reuse the manifest, even if reuse_logs
was false (the default). This was missed because the stock
Env::NewAppendableFile implementation returns false forcing the creation of a
new log.
2015-06-17 11:20:27 -07:00
Sanjay Ghemawat
251ebf5dc7 LevelDB now attempts to reuse the preceding MANIFEST and log file when re-opened.
(Based on a suggestion by cmumford.)

"open" benchmark on my workstation speeds up significantly since we
can now avoid three fdatasync calls and a compaction per open:

  Before: ~80000 microseconds
  After:    ~130 microseconds

Details:

(1) Added Options::reuse_logs (currently defaults to false) to control
new behavior.  The intention is to change the default to true after some
baking.

(2) Added Env::NewAppendableFile() whose default implementation returns
a not-supported error.

(3) VersionSet::Recovery attempts to reuse the MANIFEST from which
it is recovering.

(4) DBImpl recovery attempts to reuse the last log file and memtable.

(5) db_test.cc now tests a new configuration that sets reuse_logs to true.

(6) fault_injection_test also tests a reuse_logs==true config.

(7) Added a new recovery_test.
2014-12-11 08:13:18 -08:00
163 changed files with 8337 additions and 11321 deletions

View File

@ -1,18 +0,0 @@
# Run manually to reformat a file:
# clang-format -i --style=file <file>
# find . -iname '*.cc' -o -iname '*.h' -o -iname '*.h.in' | xargs clang-format -i --style=file
BasedOnStyle: Google
DerivePointerAlignment: false
# Public headers are in a different location in the internal Google repository.
# Order them so that when imported to the authoritative repository they will be
# in correct alphabetical order.
IncludeCategories:
- Regex: '^(<|"(benchmarks|db|helpers)/)'
Priority: 1
- Regex: '^"(leveldb)/'
Priority: 2
- Regex: '^(<|"(issues|port|table|third_party|util)/)'
Priority: 3
- Regex: '.*'
Priority: 4

View File

@ -1,102 +0,0 @@
# Copyright 2021 The LevelDB Authors. All rights reserved.
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file. See the AUTHORS file for names of contributors.
name: ci
on: [push, pull_request]
permissions:
contents: read
jobs:
build-and-test:
name: >-
CI
${{ matrix.os }}
${{ matrix.compiler }}
${{ matrix.optimized && 'release' || 'debug' }}
runs-on: ${{ matrix.os }}
strategy:
fail-fast: false
matrix:
compiler: [clang, gcc, msvc]
os: [ubuntu-latest, macos-latest, windows-latest]
optimized: [true, false]
exclude:
# MSVC only works on Windows.
- os: ubuntu-latest
compiler: msvc
- os: macos-latest
compiler: msvc
# Not testing with GCC on macOS.
- os: macos-latest
compiler: gcc
# Only testing with MSVC on Windows.
- os: windows-latest
compiler: clang
- os: windows-latest
compiler: gcc
include:
- compiler: clang
CC: clang
CXX: clang++
- compiler: gcc
CC: gcc
CXX: g++
- compiler: msvc
CC:
CXX:
env:
CMAKE_BUILD_DIR: ${{ github.workspace }}/build
CMAKE_BUILD_TYPE: ${{ matrix.optimized && 'RelWithDebInfo' || 'Debug' }}
CC: ${{ matrix.CC }}
CXX: ${{ matrix.CXX }}
BINARY_SUFFIX: ${{ startsWith(matrix.os, 'windows') && '.exe' || '' }}
BINARY_PATH: >-
${{ format(
startsWith(matrix.os, 'windows') && '{0}\build\{1}\' || '{0}/build/',
github.workspace,
matrix.optimized && 'RelWithDebInfo' || 'Debug') }}
steps:
- uses: actions/checkout@v2
with:
submodules: true
- name: Install dependencies on Linux
if: ${{ runner.os == 'Linux' }}
# libgoogle-perftools-dev is temporarily removed from the package list
# because it is currently broken on GitHub's Ubuntu 22.04.
run: |
sudo apt-get update
sudo apt-get install libkyotocabinet-dev libsnappy-dev libsqlite3-dev
- name: Generate build config
run: >-
cmake -S "${{ github.workspace }}" -B "${{ env.CMAKE_BUILD_DIR }}"
-DCMAKE_BUILD_TYPE=${{ env.CMAKE_BUILD_TYPE }}
-DCMAKE_INSTALL_PREFIX=${{ runner.temp }}/install_test/
- name: Build
run: >-
cmake --build "${{ env.CMAKE_BUILD_DIR }}"
--config "${{ env.CMAKE_BUILD_TYPE }}"
- name: Run Tests
working-directory: ${{ github.workspace }}/build
run: ctest -C "${{ env.CMAKE_BUILD_TYPE }}" --verbose
- name: Run LevelDB Benchmarks
run: ${{ env.BINARY_PATH }}db_bench${{ env.BINARY_SUFFIX }}
- name: Run SQLite Benchmarks
if: ${{ runner.os != 'Windows' }}
run: ${{ env.BINARY_PATH }}db_bench_sqlite3${{ env.BINARY_SUFFIX }}
- name: Run Kyoto Cabinet Benchmarks
if: ${{ runner.os == 'Linux' && matrix.compiler == 'clang' }}
run: ${{ env.BINARY_PATH }}db_bench_tree_db${{ env.BINARY_SUFFIX }}
- name: Test CMake installation
run: cmake --build "${{ env.CMAKE_BUILD_DIR }}" --target install

17
.gitignore vendored
View File

@ -1,8 +1,9 @@
# Editors. build_config.mk
*.sw* *.a
.vscode *.o
.DS_Store *.dylib*
*.so
# Build directory. *.so.*
build/ *_test
out/ db_bench
leveldbutil

6
.gitmodules vendored
View File

@ -1,6 +0,0 @@
[submodule "third_party/googletest"]
path = third_party/googletest
url = https://github.com/google/googletest.git
[submodule "third_party/benchmark"]
path = third_party/benchmark
url = https://github.com/google/benchmark

View File

@ -1,519 +0,0 @@
# Copyright 2017 The LevelDB Authors. All rights reserved.
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file. See the AUTHORS file for names of contributors.
cmake_minimum_required(VERSION 3.9)
# Keep the version below in sync with the one in db.h
project(leveldb VERSION 1.23.0 LANGUAGES C CXX)
# C standard can be overridden when this is used as a sub-project.
if(NOT CMAKE_C_STANDARD)
# This project can use C11, but will gracefully decay down to C89.
set(CMAKE_C_STANDARD 11)
set(CMAKE_C_STANDARD_REQUIRED OFF)
set(CMAKE_C_EXTENSIONS OFF)
endif(NOT CMAKE_C_STANDARD)
# C++ standard can be overridden when this is used as a sub-project.
if(NOT CMAKE_CXX_STANDARD)
# This project requires C++11.
set(CMAKE_CXX_STANDARD 11)
set(CMAKE_CXX_STANDARD_REQUIRED ON)
set(CMAKE_CXX_EXTENSIONS OFF)
endif(NOT CMAKE_CXX_STANDARD)
if (WIN32)
set(LEVELDB_PLATFORM_NAME LEVELDB_PLATFORM_WINDOWS)
# TODO(cmumford): Make UNICODE configurable for Windows.
add_definitions(-D_UNICODE -DUNICODE)
else (WIN32)
set(LEVELDB_PLATFORM_NAME LEVELDB_PLATFORM_POSIX)
endif (WIN32)
option(LEVELDB_BUILD_TESTS "Build LevelDB's unit tests" ON)
option(LEVELDB_BUILD_BENCHMARKS "Build LevelDB's benchmarks" ON)
option(LEVELDB_INSTALL "Install LevelDB's header and library" ON)
include(CheckIncludeFile)
check_include_file("unistd.h" HAVE_UNISTD_H)
include(CheckLibraryExists)
check_library_exists(crc32c crc32c_value "" HAVE_CRC32C)
check_library_exists(snappy snappy_compress "" HAVE_SNAPPY)
check_library_exists(zstd zstd_compress "" HAVE_ZSTD)
check_library_exists(tcmalloc malloc "" HAVE_TCMALLOC)
include(CheckCXXSymbolExists)
# Using check_cxx_symbol_exists() instead of check_c_symbol_exists() because
# we're including the header from C++, and feature detection should use the same
# compiler language that the project will use later. Principles aside, some
# versions of do not expose fdatasync() in <unistd.h> in standard C mode
# (-std=c11), but do expose the function in standard C++ mode (-std=c++11).
check_cxx_symbol_exists(fdatasync "unistd.h" HAVE_FDATASYNC)
check_cxx_symbol_exists(F_FULLFSYNC "fcntl.h" HAVE_FULLFSYNC)
check_cxx_symbol_exists(O_CLOEXEC "fcntl.h" HAVE_O_CLOEXEC)
if(CMAKE_CXX_COMPILER_ID STREQUAL "MSVC")
# Disable C++ exceptions.
string(REGEX REPLACE "/EH[a-z]+" "" CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS}")
set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} /EHs-c-")
add_definitions(-D_HAS_EXCEPTIONS=0)
# Disable RTTI.
string(REGEX REPLACE "/GR" "" CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS}")
set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} /GR-")
else(CMAKE_CXX_COMPILER_ID STREQUAL "MSVC")
# Enable strict prototype warnings for C code in clang and gcc.
if(NOT CMAKE_C_FLAGS MATCHES "-Wstrict-prototypes")
set(CMAKE_C_FLAGS "${CMAKE_C_FLAGS} -Wstrict-prototypes")
endif(NOT CMAKE_C_FLAGS MATCHES "-Wstrict-prototypes")
# Disable C++ exceptions.
string(REGEX REPLACE "-fexceptions" "" CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS}")
set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -fno-exceptions")
# Disable RTTI.
string(REGEX REPLACE "-frtti" "" CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS}")
set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -fno-rtti")
endif(CMAKE_CXX_COMPILER_ID STREQUAL "MSVC")
# Test whether -Wthread-safety is available. See
# https://clang.llvm.org/docs/ThreadSafetyAnalysis.html
include(CheckCXXCompilerFlag)
check_cxx_compiler_flag(-Wthread-safety HAVE_CLANG_THREAD_SAFETY)
# Used by googletest.
check_cxx_compiler_flag(-Wno-missing-field-initializers
LEVELDB_HAVE_NO_MISSING_FIELD_INITIALIZERS)
include(CheckCXXSourceCompiles)
# Test whether C++17 __has_include is available.
check_cxx_source_compiles("
#if defined(__has_include) && __has_include(<string>)
#include <string>
#endif
int main() { std::string str; return 0; }
" HAVE_CXX17_HAS_INCLUDE)
set(LEVELDB_PUBLIC_INCLUDE_DIR "include/leveldb")
set(LEVELDB_PORT_CONFIG_DIR "include/port")
configure_file(
"port/port_config.h.in"
"${PROJECT_BINARY_DIR}/${LEVELDB_PORT_CONFIG_DIR}/port_config.h"
)
include_directories(
"${PROJECT_BINARY_DIR}/include"
"."
)
if(BUILD_SHARED_LIBS)
# Only export LEVELDB_EXPORT symbols from the shared library.
add_compile_options(-fvisibility=hidden)
endif(BUILD_SHARED_LIBS)
# Must be included before CMAKE_INSTALL_INCLUDEDIR is used.
include(GNUInstallDirs)
add_library(leveldb "")
target_sources(leveldb
PRIVATE
"${PROJECT_BINARY_DIR}/${LEVELDB_PORT_CONFIG_DIR}/port_config.h"
"db/builder.cc"
"db/builder.h"
"db/c.cc"
"db/db_impl.cc"
"db/db_impl.h"
"db/db_iter.cc"
"db/db_iter.h"
"db/dbformat.cc"
"db/dbformat.h"
"db/dumpfile.cc"
"db/filename.cc"
"db/filename.h"
"db/log_format.h"
"db/log_reader.cc"
"db/log_reader.h"
"db/log_writer.cc"
"db/log_writer.h"
"db/memtable.cc"
"db/memtable.h"
"db/repair.cc"
"db/skiplist.h"
"db/snapshot.h"
"db/table_cache.cc"
"db/table_cache.h"
"db/version_edit.cc"
"db/version_edit.h"
"db/version_set.cc"
"db/version_set.h"
"db/write_batch_internal.h"
"db/write_batch.cc"
"port/port_stdcxx.h"
"port/port.h"
"port/thread_annotations.h"
"table/block_builder.cc"
"table/block_builder.h"
"table/block.cc"
"table/block.h"
"table/filter_block.cc"
"table/filter_block.h"
"table/format.cc"
"table/format.h"
"table/iterator_wrapper.h"
"table/iterator.cc"
"table/merger.cc"
"table/merger.h"
"table/table_builder.cc"
"table/table.cc"
"table/two_level_iterator.cc"
"table/two_level_iterator.h"
"util/arena.cc"
"util/arena.h"
"util/bloom.cc"
"util/cache.cc"
"util/coding.cc"
"util/coding.h"
"util/comparator.cc"
"util/crc32c.cc"
"util/crc32c.h"
"util/env.cc"
"util/filter_policy.cc"
"util/hash.cc"
"util/hash.h"
"util/logging.cc"
"util/logging.h"
"util/mutexlock.h"
"util/no_destructor.h"
"util/options.cc"
"util/random.h"
"util/status.cc"
# Only CMake 3.3+ supports PUBLIC sources in targets exported by "install".
$<$<VERSION_GREATER:CMAKE_VERSION,3.2>:PUBLIC>
"${LEVELDB_PUBLIC_INCLUDE_DIR}/c.h"
"${LEVELDB_PUBLIC_INCLUDE_DIR}/cache.h"
"${LEVELDB_PUBLIC_INCLUDE_DIR}/comparator.h"
"${LEVELDB_PUBLIC_INCLUDE_DIR}/db.h"
"${LEVELDB_PUBLIC_INCLUDE_DIR}/dumpfile.h"
"${LEVELDB_PUBLIC_INCLUDE_DIR}/env.h"
"${LEVELDB_PUBLIC_INCLUDE_DIR}/export.h"
"${LEVELDB_PUBLIC_INCLUDE_DIR}/filter_policy.h"
"${LEVELDB_PUBLIC_INCLUDE_DIR}/iterator.h"
"${LEVELDB_PUBLIC_INCLUDE_DIR}/options.h"
"${LEVELDB_PUBLIC_INCLUDE_DIR}/slice.h"
"${LEVELDB_PUBLIC_INCLUDE_DIR}/status.h"
"${LEVELDB_PUBLIC_INCLUDE_DIR}/table_builder.h"
"${LEVELDB_PUBLIC_INCLUDE_DIR}/table.h"
"${LEVELDB_PUBLIC_INCLUDE_DIR}/write_batch.h"
)
if (WIN32)
target_sources(leveldb
PRIVATE
"util/env_windows.cc"
"util/windows_logger.h"
)
else (WIN32)
target_sources(leveldb
PRIVATE
"util/env_posix.cc"
"util/posix_logger.h"
)
endif (WIN32)
# MemEnv is not part of the interface and could be pulled to a separate library.
target_sources(leveldb
PRIVATE
"helpers/memenv/memenv.cc"
"helpers/memenv/memenv.h"
)
target_include_directories(leveldb
PUBLIC
$<BUILD_INTERFACE:${PROJECT_SOURCE_DIR}/include>
$<INSTALL_INTERFACE:${CMAKE_INSTALL_INCLUDEDIR}>
)
set_target_properties(leveldb
PROPERTIES VERSION ${PROJECT_VERSION} SOVERSION ${PROJECT_VERSION_MAJOR})
target_compile_definitions(leveldb
PRIVATE
# Used by include/export.h when building shared libraries.
LEVELDB_COMPILE_LIBRARY
# Used by port/port.h.
${LEVELDB_PLATFORM_NAME}=1
)
if (NOT HAVE_CXX17_HAS_INCLUDE)
target_compile_definitions(leveldb
PRIVATE
LEVELDB_HAS_PORT_CONFIG_H=1
)
endif(NOT HAVE_CXX17_HAS_INCLUDE)
if(BUILD_SHARED_LIBS)
target_compile_definitions(leveldb
PUBLIC
# Used by include/export.h.
LEVELDB_SHARED_LIBRARY
)
endif(BUILD_SHARED_LIBS)
if(HAVE_CLANG_THREAD_SAFETY)
target_compile_options(leveldb
PUBLIC
-Werror -Wthread-safety)
endif(HAVE_CLANG_THREAD_SAFETY)
if(HAVE_CRC32C)
target_link_libraries(leveldb crc32c)
endif(HAVE_CRC32C)
if(HAVE_SNAPPY)
target_link_libraries(leveldb snappy)
endif(HAVE_SNAPPY)
if(HAVE_ZSTD)
target_link_libraries(leveldb zstd)
endif(HAVE_ZSTD)
if(HAVE_TCMALLOC)
target_link_libraries(leveldb tcmalloc)
endif(HAVE_TCMALLOC)
# Needed by port_stdcxx.h
find_package(Threads REQUIRED)
target_link_libraries(leveldb Threads::Threads)
add_executable(leveldbutil
"db/leveldbutil.cc"
)
target_link_libraries(leveldbutil leveldb)
if(LEVELDB_BUILD_TESTS)
enable_testing()
# Prevent overriding the parent project's compiler/linker settings on Windows.
set(gtest_force_shared_crt ON CACHE BOOL "" FORCE)
set(install_gtest OFF)
set(install_gmock OFF)
set(build_gmock ON)
# This project is tested using GoogleTest.
add_subdirectory("third_party/googletest")
# GoogleTest triggers a missing field initializers warning.
if(LEVELDB_HAVE_NO_MISSING_FIELD_INITIALIZERS)
set_property(TARGET gtest
APPEND PROPERTY COMPILE_OPTIONS -Wno-missing-field-initializers)
set_property(TARGET gmock
APPEND PROPERTY COMPILE_OPTIONS -Wno-missing-field-initializers)
endif(LEVELDB_HAVE_NO_MISSING_FIELD_INITIALIZERS)
add_executable(leveldb_tests "")
target_sources(leveldb_tests
PRIVATE
# "db/fault_injection_test.cc"
# "issues/issue178_test.cc"
# "issues/issue200_test.cc"
# "issues/issue320_test.cc"
"${PROJECT_BINARY_DIR}/${LEVELDB_PORT_CONFIG_DIR}/port_config.h"
# "util/env_test.cc"
"util/status_test.cc"
"util/no_destructor_test.cc"
"util/testutil.cc"
"util/testutil.h"
)
if(NOT BUILD_SHARED_LIBS)
target_sources(leveldb_tests
PRIVATE
"db/autocompact_test.cc"
"db/corruption_test.cc"
"db/db_test.cc"
"db/dbformat_test.cc"
"db/filename_test.cc"
"db/log_test.cc"
"db/recovery_test.cc"
"db/skiplist_test.cc"
"db/version_edit_test.cc"
"db/version_set_test.cc"
"db/write_batch_test.cc"
"helpers/memenv/memenv_test.cc"
"table/filter_block_test.cc"
"table/table_test.cc"
"util/arena_test.cc"
"util/bloom_test.cc"
"util/cache_test.cc"
"util/coding_test.cc"
"util/crc32c_test.cc"
"util/hash_test.cc"
"util/logging_test.cc"
)
endif(NOT BUILD_SHARED_LIBS)
target_link_libraries(leveldb_tests leveldb gmock gtest gtest_main)
target_compile_definitions(leveldb_tests
PRIVATE
${LEVELDB_PLATFORM_NAME}=1
)
if (NOT HAVE_CXX17_HAS_INCLUDE)
target_compile_definitions(leveldb_tests
PRIVATE
LEVELDB_HAS_PORT_CONFIG_H=1
)
endif(NOT HAVE_CXX17_HAS_INCLUDE)
add_test(NAME "leveldb_tests" COMMAND "leveldb_tests")
function(leveldb_test test_file)
get_filename_component(test_target_name "${test_file}" NAME_WE)
add_executable("${test_target_name}" "")
target_sources("${test_target_name}"
PRIVATE
"${PROJECT_BINARY_DIR}/${LEVELDB_PORT_CONFIG_DIR}/port_config.h"
"util/testutil.cc"
"util/testutil.h"
"${test_file}"
)
target_link_libraries("${test_target_name}" leveldb gmock gtest)
target_compile_definitions("${test_target_name}"
PRIVATE
${LEVELDB_PLATFORM_NAME}=1
)
if (NOT HAVE_CXX17_HAS_INCLUDE)
target_compile_definitions("${test_target_name}"
PRIVATE
LEVELDB_HAS_PORT_CONFIG_H=1
)
endif(NOT HAVE_CXX17_HAS_INCLUDE)
add_test(NAME "${test_target_name}" COMMAND "${test_target_name}")
endfunction(leveldb_test)
leveldb_test("db/c_test.c")
if(NOT BUILD_SHARED_LIBS)
# TODO(costan): This test also uses
# "util/env_{posix|windows}_test_helper.h"
if (WIN32)
leveldb_test("util/env_windows_test.cc")
else (WIN32)
leveldb_test("util/env_posix_test.cc")
endif (WIN32)
endif(NOT BUILD_SHARED_LIBS)
endif(LEVELDB_BUILD_TESTS)
if(LEVELDB_BUILD_BENCHMARKS)
# This project uses Google benchmark for benchmarking.
set(BENCHMARK_ENABLE_TESTING OFF CACHE BOOL "" FORCE)
set(BENCHMARK_ENABLE_EXCEPTIONS OFF CACHE BOOL "" FORCE)
add_subdirectory("third_party/benchmark")
function(leveldb_benchmark bench_file)
get_filename_component(bench_target_name "${bench_file}" NAME_WE)
add_executable("${bench_target_name}" "")
target_sources("${bench_target_name}"
PRIVATE
"${PROJECT_BINARY_DIR}/${LEVELDB_PORT_CONFIG_DIR}/port_config.h"
"util/histogram.cc"
"util/histogram.h"
"util/testutil.cc"
"util/testutil.h"
"${bench_file}"
)
target_link_libraries("${bench_target_name}" leveldb gmock gtest benchmark)
target_compile_definitions("${bench_target_name}"
PRIVATE
${LEVELDB_PLATFORM_NAME}=1
)
if (NOT HAVE_CXX17_HAS_INCLUDE)
target_compile_definitions("${bench_target_name}"
PRIVATE
LEVELDB_HAS_PORT_CONFIG_H=1
)
endif(NOT HAVE_CXX17_HAS_INCLUDE)
endfunction(leveldb_benchmark)
if(NOT BUILD_SHARED_LIBS)
leveldb_benchmark("benchmarks/db_bench.cc")
endif(NOT BUILD_SHARED_LIBS)
check_library_exists(sqlite3 sqlite3_open "" HAVE_SQLITE3)
if(HAVE_SQLITE3)
leveldb_benchmark("benchmarks/db_bench_sqlite3.cc")
target_link_libraries(db_bench_sqlite3 sqlite3)
endif(HAVE_SQLITE3)
# check_library_exists is insufficient here because the library names have
# different manglings when compiled with clang or gcc, at least when installed
# with Homebrew on Mac.
set(OLD_CMAKE_REQURED_LIBRARIES ${CMAKE_REQUIRED_LIBRARIES})
list(APPEND CMAKE_REQUIRED_LIBRARIES kyotocabinet)
check_cxx_source_compiles("
#include <kcpolydb.h>
int main() {
kyotocabinet::TreeDB* db = new kyotocabinet::TreeDB();
delete db;
return 0;
}
" HAVE_KYOTOCABINET)
set(CMAKE_REQUIRED_LIBRARIES ${OLD_CMAKE_REQURED_LIBRARIES})
if(HAVE_KYOTOCABINET)
leveldb_benchmark("benchmarks/db_bench_tree_db.cc")
target_link_libraries(db_bench_tree_db kyotocabinet)
endif(HAVE_KYOTOCABINET)
endif(LEVELDB_BUILD_BENCHMARKS)
if(LEVELDB_INSTALL)
install(TARGETS leveldb
EXPORT leveldbTargets
RUNTIME DESTINATION ${CMAKE_INSTALL_BINDIR}
LIBRARY DESTINATION ${CMAKE_INSTALL_LIBDIR}
ARCHIVE DESTINATION ${CMAKE_INSTALL_LIBDIR}
)
install(
FILES
"${LEVELDB_PUBLIC_INCLUDE_DIR}/c.h"
"${LEVELDB_PUBLIC_INCLUDE_DIR}/cache.h"
"${LEVELDB_PUBLIC_INCLUDE_DIR}/comparator.h"
"${LEVELDB_PUBLIC_INCLUDE_DIR}/db.h"
"${LEVELDB_PUBLIC_INCLUDE_DIR}/dumpfile.h"
"${LEVELDB_PUBLIC_INCLUDE_DIR}/env.h"
"${LEVELDB_PUBLIC_INCLUDE_DIR}/export.h"
"${LEVELDB_PUBLIC_INCLUDE_DIR}/filter_policy.h"
"${LEVELDB_PUBLIC_INCLUDE_DIR}/iterator.h"
"${LEVELDB_PUBLIC_INCLUDE_DIR}/options.h"
"${LEVELDB_PUBLIC_INCLUDE_DIR}/slice.h"
"${LEVELDB_PUBLIC_INCLUDE_DIR}/status.h"
"${LEVELDB_PUBLIC_INCLUDE_DIR}/table_builder.h"
"${LEVELDB_PUBLIC_INCLUDE_DIR}/table.h"
"${LEVELDB_PUBLIC_INCLUDE_DIR}/write_batch.h"
DESTINATION "${CMAKE_INSTALL_INCLUDEDIR}/leveldb"
)
include(CMakePackageConfigHelpers)
configure_package_config_file(
"cmake/${PROJECT_NAME}Config.cmake.in"
"${PROJECT_BINARY_DIR}/cmake/${PROJECT_NAME}Config.cmake"
INSTALL_DESTINATION "${CMAKE_INSTALL_LIBDIR}/cmake/${PROJECT_NAME}"
)
write_basic_package_version_file(
"${PROJECT_BINARY_DIR}/cmake/${PROJECT_NAME}ConfigVersion.cmake"
COMPATIBILITY SameMajorVersion
)
install(
EXPORT leveldbTargets
NAMESPACE leveldb::
DESTINATION "${CMAKE_INSTALL_LIBDIR}/cmake/${PROJECT_NAME}"
)
install(
FILES
"${PROJECT_BINARY_DIR}/cmake/${PROJECT_NAME}Config.cmake"
"${PROJECT_BINARY_DIR}/cmake/${PROJECT_NAME}ConfigVersion.cmake"
DESTINATION "${CMAKE_INSTALL_LIBDIR}/cmake/${PROJECT_NAME}"
)
endif(LEVELDB_INSTALL)

View File

@ -1,31 +1,36 @@
# How to Contribute # Contributing
We'd love to accept your patches and contributions to this project. There are We'd love to accept your code patches! However, before we can take them, we
just a few small guidelines you need to follow. have to jump a couple of legal hurdles.
## Contributor License Agreement ## Contributor License Agreements
Contributions to this project must be accompanied by a Contributor License Please fill out either the individual or corporate Contributor License
Agreement. You (or your employer) retain the copyright to your contribution; Agreement as appropriate.
this simply gives us permission to use and redistribute your contributions as
part of the project. Head over to <https://cla.developers.google.com/> to see
your current agreements on file or to sign a new one.
You generally only need to submit a CLA once, so if you've already submitted one * If you are an individual writing original source code and you're sure you
(even if it was for a different project), you probably don't need to do it own the intellectual property, then sign an [individual CLA](https://developers.google.com/open-source/cla/individual).
again. * If you work for a company that wants to allow you to contribute your work,
then sign a [corporate CLA](https://developers.google.com/open-source/cla/corporate).
## Code Reviews Follow either of the two links above to access the appropriate CLA and
instructions for how to sign and return it.
All submissions, including submissions by project members, require review. We ## Submitting a Patch
use GitHub pull requests for this purpose. Consult
[GitHub Help](https://help.github.com/articles/about-pull-requests/) for more
information on using pull requests.
See [the README](README.md#contributing-to-the-leveldb-project) for areas 1. Sign the contributors license agreement above.
where we are likely to accept external contributions. 2. Decide which code you want to submit. A submission should be a set of changes
that addresses one issue in the [issue tracker](https://github.com/google/leveldb/issues).
Please don't mix more than one logical change per submission, because it makes
the history hard to follow. If you want to make a change
(e.g. add a sample or feature) that doesn't have a corresponding issue in the
issue tracker, please create one.
3. **Submitting**: When you are ready to submit, send us a Pull Request. Be
sure to include the issue number you fixed and the name you used to sign
the CLA.
## Community Guidelines ## Writing Code ##
This project follows [Google's Open Source Community If your contribution contains code, please make sure that it follows
Guidelines](https://opensource.google/conduct/). [the style guide](http://google-styleguide.googlecode.com/svn/trunk/cppguide.xml).
Otherwise we will have to ask you to make changes, and that's no fun for anyone.

233
Makefile Normal file
View File

@ -0,0 +1,233 @@
# Copyright (c) 2011 The LevelDB Authors. All rights reserved.
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file. See the AUTHORS file for names of contributors.
#-----------------------------------------------
# Uncomment exactly one of the lines labelled (A), (B), and (C) below
# to switch between compilation modes.
# (A) Production use (optimized mode)
OPT ?= -O2 -DNDEBUG
# (B) Debug mode, w/ full line-level debugging symbols
# OPT ?= -g2
# (C) Profiling mode: opt, but w/debugging symbols
# OPT ?= -O2 -g2 -DNDEBUG
#-----------------------------------------------
# detect what platform we're building on
$(shell CC="$(CC)" CXX="$(CXX)" TARGET_OS="$(TARGET_OS)" \
./build_detect_platform build_config.mk ./)
# this file is generated by the previous line to set build flags and sources
include build_config.mk
CFLAGS += -I. -I./include $(PLATFORM_CCFLAGS) $(OPT)
CXXFLAGS += -I. -I./include $(PLATFORM_CXXFLAGS) $(OPT)
LDFLAGS += $(PLATFORM_LDFLAGS)
LIBS += $(PLATFORM_LIBS)
LIBOBJECTS = $(SOURCES:.cc=.o)
MEMENVOBJECTS = $(MEMENV_SOURCES:.cc=.o)
TESTUTIL = ./util/testutil.o
TESTHARNESS = ./util/testharness.o $(TESTUTIL)
# Note: iOS should probably be using libtool, not ar.
ifeq ($(PLATFORM), IOS)
AR=xcrun ar
endif
TESTS = \
arena_test \
autocompact_test \
bloom_test \
c_test \
cache_test \
coding_test \
corruption_test \
crc32c_test \
db_test \
dbformat_test \
env_test \
fault_injection_test \
filename_test \
filter_block_test \
hash_test \
issue178_test \
issue200_test \
log_test \
memenv_test \
recovery_test \
skiplist_test \
table_test \
version_edit_test \
version_set_test \
write_batch_test
PROGRAMS = db_bench leveldbutil $(TESTS)
BENCHMARKS = db_bench_sqlite3 db_bench_tree_db
LIBRARY = libleveldb.a
MEMENVLIBRARY = libmemenv.a
default: all
# Should we build shared libraries?
ifneq ($(PLATFORM_SHARED_EXT),)
ifneq ($(PLATFORM_SHARED_VERSIONED),true)
SHARED1 = libleveldb.$(PLATFORM_SHARED_EXT)
SHARED2 = $(SHARED1)
SHARED3 = $(SHARED1)
SHARED = $(SHARED1)
else
# Update db.h if you change these.
SHARED_MAJOR = 1
SHARED_MINOR = 18
SHARED1 = libleveldb.$(PLATFORM_SHARED_EXT)
SHARED2 = $(SHARED1).$(SHARED_MAJOR)
SHARED3 = $(SHARED1).$(SHARED_MAJOR).$(SHARED_MINOR)
SHARED = $(SHARED1) $(SHARED2) $(SHARED3)
$(SHARED1): $(SHARED3)
ln -fs $(SHARED3) $(SHARED1)
$(SHARED2): $(SHARED3)
ln -fs $(SHARED3) $(SHARED2)
endif
$(SHARED3):
$(CXX) $(LDFLAGS) $(PLATFORM_SHARED_LDFLAGS)$(SHARED2) $(CXXFLAGS) $(PLATFORM_SHARED_CFLAGS) $(SOURCES) -o $(SHARED3) $(LIBS)
endif # PLATFORM_SHARED_EXT
all: $(SHARED) $(LIBRARY)
check: all $(PROGRAMS) $(TESTS)
for t in $(TESTS); do echo "***** Running $$t"; ./$$t || exit 1; done
clean:
-rm -f $(PROGRAMS) $(BENCHMARKS) $(LIBRARY) $(SHARED) $(MEMENVLIBRARY) */*.o */*/*.o ios-x86/*/*.o ios-arm/*/*.o build_config.mk
-rm -rf ios-x86/* ios-arm/*
$(LIBRARY): $(LIBOBJECTS)
rm -f $@
$(AR) -rs $@ $(LIBOBJECTS)
db_bench: db/db_bench.o $(LIBOBJECTS) $(TESTUTIL)
$(CXX) $(LDFLAGS) db/db_bench.o $(LIBOBJECTS) $(TESTUTIL) -o $@ $(LIBS)
db_bench_sqlite3: doc/bench/db_bench_sqlite3.o $(LIBOBJECTS) $(TESTUTIL)
$(CXX) $(LDFLAGS) doc/bench/db_bench_sqlite3.o $(LIBOBJECTS) $(TESTUTIL) -o $@ -lsqlite3 $(LIBS)
db_bench_tree_db: doc/bench/db_bench_tree_db.o $(LIBOBJECTS) $(TESTUTIL)
$(CXX) $(LDFLAGS) doc/bench/db_bench_tree_db.o $(LIBOBJECTS) $(TESTUTIL) -o $@ -lkyotocabinet $(LIBS)
leveldbutil: db/leveldb_main.o $(LIBOBJECTS)
$(CXX) $(LDFLAGS) db/leveldb_main.o $(LIBOBJECTS) -o $@ $(LIBS)
arena_test: util/arena_test.o $(LIBOBJECTS) $(TESTHARNESS)
$(CXX) $(LDFLAGS) util/arena_test.o $(LIBOBJECTS) $(TESTHARNESS) -o $@ $(LIBS)
autocompact_test: db/autocompact_test.o $(LIBOBJECTS) $(TESTHARNESS)
$(CXX) $(LDFLAGS) db/autocompact_test.o $(LIBOBJECTS) $(TESTHARNESS) -o $@ $(LIBS)
bloom_test: util/bloom_test.o $(LIBOBJECTS) $(TESTHARNESS)
$(CXX) $(LDFLAGS) util/bloom_test.o $(LIBOBJECTS) $(TESTHARNESS) -o $@ $(LIBS)
c_test: db/c_test.o $(LIBOBJECTS) $(TESTHARNESS)
$(CXX) $(LDFLAGS) db/c_test.o $(LIBOBJECTS) $(TESTHARNESS) -o $@ $(LIBS)
cache_test: util/cache_test.o $(LIBOBJECTS) $(TESTHARNESS)
$(CXX) $(LDFLAGS) util/cache_test.o $(LIBOBJECTS) $(TESTHARNESS) -o $@ $(LIBS)
coding_test: util/coding_test.o $(LIBOBJECTS) $(TESTHARNESS)
$(CXX) $(LDFLAGS) util/coding_test.o $(LIBOBJECTS) $(TESTHARNESS) -o $@ $(LIBS)
corruption_test: db/corruption_test.o $(LIBOBJECTS) $(TESTHARNESS)
$(CXX) $(LDFLAGS) db/corruption_test.o $(LIBOBJECTS) $(TESTHARNESS) -o $@ $(LIBS)
crc32c_test: util/crc32c_test.o $(LIBOBJECTS) $(TESTHARNESS)
$(CXX) $(LDFLAGS) util/crc32c_test.o $(LIBOBJECTS) $(TESTHARNESS) -o $@ $(LIBS)
db_test: db/db_test.o $(LIBOBJECTS) $(TESTHARNESS)
$(CXX) $(LDFLAGS) db/db_test.o $(LIBOBJECTS) $(TESTHARNESS) -o $@ $(LIBS)
dbformat_test: db/dbformat_test.o $(LIBOBJECTS) $(TESTHARNESS)
$(CXX) $(LDFLAGS) db/dbformat_test.o $(LIBOBJECTS) $(TESTHARNESS) -o $@ $(LIBS)
env_test: util/env_test.o $(LIBOBJECTS) $(TESTHARNESS)
$(CXX) $(LDFLAGS) util/env_test.o $(LIBOBJECTS) $(TESTHARNESS) -o $@ $(LIBS)
fault_injection_test: db/fault_injection_test.o $(LIBOBJECTS) $(TESTHARNESS)
$(CXX) $(LDFLAGS) db/fault_injection_test.o $(LIBOBJECTS) $(TESTHARNESS) -o $@ $(LIBS)
filename_test: db/filename_test.o $(LIBOBJECTS) $(TESTHARNESS)
$(CXX) $(LDFLAGS) db/filename_test.o $(LIBOBJECTS) $(TESTHARNESS) -o $@ $(LIBS)
filter_block_test: table/filter_block_test.o $(LIBOBJECTS) $(TESTHARNESS)
$(CXX) $(LDFLAGS) table/filter_block_test.o $(LIBOBJECTS) $(TESTHARNESS) -o $@ $(LIBS)
hash_test: util/hash_test.o $(LIBOBJECTS) $(TESTHARNESS)
$(CXX) $(LDFLAGS) util/hash_test.o $(LIBOBJECTS) $(TESTHARNESS) -o $@ $(LIBS)
issue178_test: issues/issue178_test.o $(LIBOBJECTS) $(TESTHARNESS)
$(CXX) $(LDFLAGS) issues/issue178_test.o $(LIBOBJECTS) $(TESTHARNESS) -o $@ $(LIBS)
issue200_test: issues/issue200_test.o $(LIBOBJECTS) $(TESTHARNESS)
$(CXX) $(LDFLAGS) issues/issue200_test.o $(LIBOBJECTS) $(TESTHARNESS) -o $@ $(LIBS)
log_test: db/log_test.o $(LIBOBJECTS) $(TESTHARNESS)
$(CXX) $(LDFLAGS) db/log_test.o $(LIBOBJECTS) $(TESTHARNESS) -o $@ $(LIBS)
recovery_test: db/recovery_test.o $(LIBOBJECTS) $(TESTHARNESS)
$(CXX) $(LDFLAGS) db/recovery_test.o $(LIBOBJECTS) $(TESTHARNESS) -o $@ $(LIBS)
table_test: table/table_test.o $(LIBOBJECTS) $(TESTHARNESS)
$(CXX) $(LDFLAGS) table/table_test.o $(LIBOBJECTS) $(TESTHARNESS) -o $@ $(LIBS)
skiplist_test: db/skiplist_test.o $(LIBOBJECTS) $(TESTHARNESS)
$(CXX) $(LDFLAGS) db/skiplist_test.o $(LIBOBJECTS) $(TESTHARNESS) -o $@ $(LIBS)
version_edit_test: db/version_edit_test.o $(LIBOBJECTS) $(TESTHARNESS)
$(CXX) $(LDFLAGS) db/version_edit_test.o $(LIBOBJECTS) $(TESTHARNESS) -o $@ $(LIBS)
version_set_test: db/version_set_test.o $(LIBOBJECTS) $(TESTHARNESS)
$(CXX) $(LDFLAGS) db/version_set_test.o $(LIBOBJECTS) $(TESTHARNESS) -o $@ $(LIBS)
write_batch_test: db/write_batch_test.o $(LIBOBJECTS) $(TESTHARNESS)
$(CXX) $(LDFLAGS) db/write_batch_test.o $(LIBOBJECTS) $(TESTHARNESS) -o $@ $(LIBS)
$(MEMENVLIBRARY) : $(MEMENVOBJECTS)
rm -f $@
$(AR) -rs $@ $(MEMENVOBJECTS)
memenv_test : helpers/memenv/memenv_test.o $(MEMENVLIBRARY) $(LIBRARY) $(TESTHARNESS)
$(CXX) $(LDFLAGS) helpers/memenv/memenv_test.o $(MEMENVLIBRARY) $(LIBRARY) $(TESTHARNESS) -o $@ $(LIBS)
ifeq ($(PLATFORM), IOS)
# For iOS, create universal object files to be used on both the simulator and
# a device.
SIMULATORSDK=$(shell xcrun -sdk iphonesimulator --show-sdk-path)
DEVICESDK=$(shell xcrun -sdk iphoneos --show-sdk-path)
IOSARCH=-arch armv6 -arch armv7 -arch armv7s -arch arm64
.cc.o:
mkdir -p ios-x86/$(dir $@)
xcrun -sdk iphonesimulator $(CXX) $(CXXFLAGS) -isysroot "$(SIMULATORSDK)" -arch i686 -arch x86_64 -c $< -o ios-x86/$@
mkdir -p ios-arm/$(dir $@)
xcrun -sdk iphoneos $(CXX) $(CXXFLAGS) -isysroot "$(DEVICESDK)" $(IOSARCH) -c $< -o ios-arm/$@
xcrun lipo ios-x86/$@ ios-arm/$@ -create -output $@
.c.o:
mkdir -p ios-x86/$(dir $@)
xcrun -sdk iphonesimulator $(CC) $(CFLAGS) -isysroot "$(SIMULATORSDK)" -arch i686 -arch x86_64 -c $< -o ios-x86/$@
mkdir -p ios-arm/$(dir $@)
xcrun -sdk iphoneos $(CC) $(CFLAGS) -isysroot "$(DEVICESDK)" $(IOSARCH) -c $< -o ios-arm/$@
xcrun lipo ios-x86/$@ ios-arm/$@ -create -output $@
else
.cc.o:
$(CXX) $(CXXFLAGS) -c $< -o $@
.c.o:
$(CC) $(CFLAGS) -c $< -o $@
endif

158
README.md
View File

@ -1,16 +1,8 @@
LevelDB is a fast key-value storage library written at Google that provides an ordered mapping from string keys to string values. **LevelDB is a fast key-value storage library written at Google that provides an ordered mapping from string keys to string values.**
> **This repository is receiving very limited maintenance. We will only review the following types of changes.**
>
> * Fixes for critical bugs, such as data loss or memory corruption
> * Changes absolutely needed by internally supported leveldb clients. These typically fix breakage introduced by a language/standard library/OS update
[![ci](https://github.com/google/leveldb/actions/workflows/build.yml/badge.svg)](https://github.com/google/leveldb/actions/workflows/build.yml)
Authors: Sanjay Ghemawat (sanjay@google.com) and Jeff Dean (jeff@google.com) Authors: Sanjay Ghemawat (sanjay@google.com) and Jeff Dean (jeff@google.com)
# Features # Features
* Keys and values are arbitrary byte arrays. * Keys and values are arbitrary byte arrays.
* Data is stored sorted by key. * Data is stored sorted by key.
* Callers can provide a custom comparison function to override the sort order. * Callers can provide a custom comparison function to override the sort order.
@ -18,115 +10,16 @@ Authors: Sanjay Ghemawat (sanjay@google.com) and Jeff Dean (jeff@google.com)
* Multiple changes can be made in one atomic batch. * Multiple changes can be made in one atomic batch.
* Users can create a transient snapshot to get a consistent view of data. * Users can create a transient snapshot to get a consistent view of data.
* Forward and backward iteration is supported over the data. * Forward and backward iteration is supported over the data.
* Data is automatically compressed using the [Snappy compression library](https://google.github.io/snappy/), but [Zstd compression](https://facebook.github.io/zstd/) is also supported. * Data is automatically compressed using the [Snappy compression library](http://code.google.com/p/snappy).
* External activity (file system operations etc.) is relayed through a virtual interface so users can customize the operating system interactions. * External activity (file system operations etc.) is relayed through a virtual interface so users can customize the operating system interactions.
* [Detailed documentation](http://htmlpreview.github.io/?https://github.com/google/leveldb/blob/master/doc/index.html) about how to use the library is included with the source code.
# Documentation
[LevelDB library documentation](https://github.com/google/leveldb/blob/main/doc/index.md) is online and bundled with the source code.
# Limitations # Limitations
* This is not a SQL database. It does not have a relational data model, it does not support SQL queries, and it has no support for indexes. * This is not a SQL database. It does not have a relational data model, it does not support SQL queries, and it has no support for indexes.
* Only a single process (possibly multi-threaded) can access a particular database at a time. * Only a single process (possibly multi-threaded) can access a particular database at a time.
* There is no client-server support builtin to the library. An application that needs such support will have to wrap their own server around the library. * There is no client-server support builtin to the library. An application that needs such support will have to wrap their own server around the library.
# Getting the Source
```bash
git clone --recurse-submodules https://github.com/google/leveldb.git
```
# Building
This project supports [CMake](https://cmake.org/) out of the box.
### Build for POSIX
Quick start:
```bash
mkdir -p build && cd build
cmake -DCMAKE_BUILD_TYPE=Release .. && cmake --build .
```
### Building for Windows
First generate the Visual Studio 2017 project/solution files:
```cmd
mkdir build
cd build
cmake -G "Visual Studio 15" ..
```
The default default will build for x86. For 64-bit run:
```cmd
cmake -G "Visual Studio 15 Win64" ..
```
To compile the Windows solution from the command-line:
```cmd
devenv /build Debug leveldb.sln
```
or open leveldb.sln in Visual Studio and build from within.
Please see the CMake documentation and `CMakeLists.txt` for more advanced usage.
# Contributing to the leveldb Project
> **This repository is receiving very limited maintenance. We will only review the following types of changes.**
>
> * Bug fixes
> * Changes absolutely needed by internally supported leveldb clients. These typically fix breakage introduced by a language/standard library/OS update
The leveldb project welcomes contributions. leveldb's primary goal is to be
a reliable and fast key/value store. Changes that are in line with the
features/limitations outlined above, and meet the requirements below,
will be considered.
Contribution requirements:
1. **Tested platforms only**. We _generally_ will only accept changes for
platforms that are compiled and tested. This means POSIX (for Linux and
macOS) or Windows. Very small changes will sometimes be accepted, but
consider that more of an exception than the rule.
2. **Stable API**. We strive very hard to maintain a stable API. Changes that
require changes for projects using leveldb _might_ be rejected without
sufficient benefit to the project.
3. **Tests**: All changes must be accompanied by a new (or changed) test, or
a sufficient explanation as to why a new (or changed) test is not required.
4. **Consistent Style**: This project conforms to the
[Google C++ Style Guide](https://google.github.io/styleguide/cppguide.html).
To ensure your changes are properly formatted please run:
```
clang-format -i --style=file <file>
```
We are unlikely to accept contributions to the build configuration files, such
as `CMakeLists.txt`. We are focused on maintaining a build configuration that
allows us to test that the project works in a few supported configurations
inside Google. We are not currently interested in supporting other requirements,
such as different operating systems, compilers, or build systems.
## Submitting a Pull Request
Before any pull request will be accepted the author must first sign a
Contributor License Agreement (CLA) at https://cla.developers.google.com/.
In order to keep the commit timeline linear
[squash](https://git-scm.com/book/en/v2/Git-Tools-Rewriting-History#Squashing-Commits)
your changes down to a single commit and [rebase](https://git-scm.com/docs/git-rebase)
on google/leveldb/main. This keeps the commit timeline linear and more easily sync'ed
with the internal repository at Google. More information at GitHub's
[About Git rebase](https://help.github.com/articles/about-git-rebase/) page.
# Performance # Performance
Here is a performance report (with explanations) from the run of the Here is a performance report (with explanations) from the run of the
@ -185,62 +78,61 @@ by the one or two disk seeks needed to fetch the data from disk.
Write performance will be mostly unaffected by whether or not the Write performance will be mostly unaffected by whether or not the
working set fits in memory. working set fits in memory.
readrandom : 16.677 micros/op; (approximately 60,000 reads per second) readrandom : 16.677 micros/op; (approximately 60,000 reads per second)
readseq : 0.476 micros/op; 232.3 MB/s readseq : 0.476 micros/op; 232.3 MB/s
readreverse : 0.724 micros/op; 152.9 MB/s readreverse : 0.724 micros/op; 152.9 MB/s
LevelDB compacts its underlying storage data in the background to LevelDB compacts its underlying storage data in the background to
improve read performance. The results listed above were done improve read performance. The results listed above were done
immediately after a lot of random writes. The results after immediately after a lot of random writes. The results after
compactions (which are usually triggered automatically) are better. compactions (which are usually triggered automatically) are better.
readrandom : 11.602 micros/op; (approximately 85,000 reads per second) readrandom : 11.602 micros/op; (approximately 85,000 reads per second)
readseq : 0.423 micros/op; 261.8 MB/s readseq : 0.423 micros/op; 261.8 MB/s
readreverse : 0.663 micros/op; 166.9 MB/s readreverse : 0.663 micros/op; 166.9 MB/s
Some of the high cost of reads comes from repeated decompression of blocks Some of the high cost of reads comes from repeated decompression of blocks
read from disk. If we supply enough cache to the leveldb so it can hold the read from disk. If we supply enough cache to the leveldb so it can hold the
uncompressed blocks in memory, the read performance improves again: uncompressed blocks in memory, the read performance improves again:
readrandom : 9.775 micros/op; (approximately 100,000 reads per second before compaction) readrandom : 9.775 micros/op; (approximately 100,000 reads per second before compaction)
readrandom : 5.215 micros/op; (approximately 190,000 reads per second after compaction) readrandom : 5.215 micros/op; (approximately 190,000 reads per second after compaction)
## Repository contents ## Repository contents
See [doc/index.md](doc/index.md) for more explanation. See See doc/index.html for more explanation. See doc/impl.html for a brief overview of the implementation.
[doc/impl.md](doc/impl.md) for a brief overview of the implementation.
The public interface is in include/leveldb/*.h. Callers should not include or The public interface is in include/*.h. Callers should not include or
rely on the details of any other header files in this package. Those rely on the details of any other header files in this package. Those
internal APIs may be changed without warning. internal APIs may be changed without warning.
Guide to header files: Guide to header files:
* **include/leveldb/db.h**: Main interface to the DB: Start here. * **include/db.h**: Main interface to the DB: Start here
* **include/leveldb/options.h**: Control over the behavior of an entire database, * **include/options.h**: Control over the behavior of an entire database,
and also control over the behavior of individual reads and writes. and also control over the behavior of individual reads and writes.
* **include/leveldb/comparator.h**: Abstraction for user-specified comparison function. * **include/comparator.h**: Abstraction for user-specified comparison function.
If you want just bytewise comparison of keys, you can use the default If you want just bytewise comparison of keys, you can use the default
comparator, but clients can write their own comparator implementations if they comparator, but clients can write their own comparator implementations if they
want custom ordering (e.g. to handle different character encodings, etc.). want custom ordering (e.g. to handle different character encodings, etc.)
* **include/leveldb/iterator.h**: Interface for iterating over data. You can get * **include/iterator.h**: Interface for iterating over data. You can get
an iterator from a DB object. an iterator from a DB object.
* **include/leveldb/write_batch.h**: Interface for atomically applying multiple * **include/write_batch.h**: Interface for atomically applying multiple
updates to a database. updates to a database.
* **include/leveldb/slice.h**: A simple module for maintaining a pointer and a * **include/slice.h**: A simple module for maintaining a pointer and a
length into some other byte array. length into some other byte array.
* **include/leveldb/status.h**: Status is returned from many of the public interfaces * **include/status.h**: Status is returned from many of the public interfaces
and is used to report success and various kinds of errors. and is used to report success and various kinds of errors.
* **include/leveldb/env.h**: * **include/env.h**:
Abstraction of the OS environment. A posix implementation of this interface is Abstraction of the OS environment. A posix implementation of this interface is
in util/env_posix.cc. in util/env_posix.cc
* **include/leveldb/table.h, include/leveldb/table_builder.h**: Lower-level modules that most * **include/table.h, include/table_builder.h**: Lower-level modules that most
clients probably won't use directly. clients probably won't use directly

View File

@ -1,92 +0,0 @@
// Copyright (c) 2019 The LevelDB Authors. All rights reserved.
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file. See the AUTHORS file for names of contributors.
#include <cinttypes>
#include <cstdio>
#include <string>
#include "gtest/gtest.h"
#include "benchmark/benchmark.h"
#include "db/version_set.h"
#include "leveldb/comparator.h"
#include "leveldb/db.h"
#include "leveldb/env.h"
#include "leveldb/options.h"
#include "port/port.h"
#include "util/mutexlock.h"
#include "util/testutil.h"
namespace leveldb {
namespace {
std::string MakeKey(unsigned int num) {
char buf[30];
std::snprintf(buf, sizeof(buf), "%016u", num);
return std::string(buf);
}
void BM_LogAndApply(benchmark::State& state) {
const int num_base_files = state.range(0);
std::string dbname = testing::TempDir() + "leveldb_test_benchmark";
DestroyDB(dbname, Options());
DB* db = nullptr;
Options opts;
opts.create_if_missing = true;
Status s = DB::Open(opts, dbname, &db);
ASSERT_LEVELDB_OK(s);
ASSERT_TRUE(db != nullptr);
delete db;
db = nullptr;
Env* env = Env::Default();
port::Mutex mu;
MutexLock l(&mu);
InternalKeyComparator cmp(BytewiseComparator());
Options options;
VersionSet vset(dbname, &options, nullptr, &cmp);
bool save_manifest;
ASSERT_LEVELDB_OK(vset.Recover(&save_manifest));
VersionEdit vbase;
uint64_t fnum = 1;
for (int i = 0; i < num_base_files; i++) {
InternalKey start(MakeKey(2 * fnum), 1, kTypeValue);
InternalKey limit(MakeKey(2 * fnum + 1), 1, kTypeDeletion);
vbase.AddFile(2, fnum++, 1 /* file size */, start, limit);
}
ASSERT_LEVELDB_OK(vset.LogAndApply(&vbase, &mu));
uint64_t start_micros = env->NowMicros();
for (auto st : state) {
VersionEdit vedit;
vedit.RemoveFile(2, fnum);
InternalKey start(MakeKey(2 * fnum), 1, kTypeValue);
InternalKey limit(MakeKey(2 * fnum + 1), 1, kTypeDeletion);
vedit.AddFile(2, fnum++, 1 /* file size */, start, limit);
vset.LogAndApply(&vedit, &mu);
}
uint64_t stop_micros = env->NowMicros();
unsigned int us = stop_micros - start_micros;
char buf[16];
std::snprintf(buf, sizeof(buf), "%d", num_base_files);
std::fprintf(stderr,
"BM_LogAndApply/%-6s %8" PRIu64
" iters : %9u us (%7.0f us / iter)\n",
buf, state.iterations(), us, ((float)us) / state.iterations());
}
BENCHMARK(BM_LogAndApply)->Arg(1)->Arg(100)->Arg(10000)->Arg(100000);
} // namespace
} // namespace leveldb
BENCHMARK_MAIN();

228
build_detect_platform Executable file
View File

@ -0,0 +1,228 @@
#!/bin/sh
#
# Detects OS we're compiling on and outputs a file specified by the first
# argument, which in turn gets read while processing Makefile.
#
# The output will set the following variables:
# CC C Compiler path
# CXX C++ Compiler path
# PLATFORM_LDFLAGS Linker flags
# PLATFORM_LIBS Libraries flags
# PLATFORM_SHARED_EXT Extension for shared libraries
# PLATFORM_SHARED_LDFLAGS Flags for building shared library
# This flag is embedded just before the name
# of the shared library without intervening spaces
# PLATFORM_SHARED_CFLAGS Flags for compiling objects for shared library
# PLATFORM_CCFLAGS C compiler flags
# PLATFORM_CXXFLAGS C++ compiler flags. Will contain:
# PLATFORM_SHARED_VERSIONED Set to 'true' if platform supports versioned
# shared libraries, empty otherwise.
#
# The PLATFORM_CCFLAGS and PLATFORM_CXXFLAGS might include the following:
#
# -DLEVELDB_ATOMIC_PRESENT if <atomic> is present
# -DLEVELDB_PLATFORM_POSIX for Posix-based platforms
# -DSNAPPY if the Snappy library is present
#
OUTPUT=$1
PREFIX=$2
if test -z "$OUTPUT" || test -z "$PREFIX"; then
echo "usage: $0 <output-filename> <directory_prefix>" >&2
exit 1
fi
# Delete existing output, if it exists
rm -f $OUTPUT
touch $OUTPUT
if test -z "$CC"; then
CC=cc
fi
if test -z "$CXX"; then
CXX=g++
fi
if test -z "$TMPDIR"; then
TMPDIR=/tmp
fi
# Detect OS
if test -z "$TARGET_OS"; then
TARGET_OS=`uname -s`
fi
COMMON_FLAGS=
CROSS_COMPILE=
PLATFORM_CCFLAGS=
PLATFORM_CXXFLAGS=
PLATFORM_LDFLAGS=
PLATFORM_LIBS=
PLATFORM_SHARED_EXT="so"
PLATFORM_SHARED_LDFLAGS="-shared -Wl,-soname -Wl,"
PLATFORM_SHARED_CFLAGS="-fPIC"
PLATFORM_SHARED_VERSIONED=true
MEMCMP_FLAG=
if [ "$CXX" = "g++" ]; then
# Use libc's memcmp instead of GCC's memcmp. This results in ~40%
# performance improvement on readrandom under gcc 4.4.3 on Linux/x86.
MEMCMP_FLAG="-fno-builtin-memcmp"
fi
case "$TARGET_OS" in
CYGWIN_*)
PLATFORM=OS_LINUX
COMMON_FLAGS="$MEMCMP_FLAG -lpthread -DOS_LINUX -DCYGWIN"
PLATFORM_LDFLAGS="-lpthread"
PORT_FILE=port/port_posix.cc
;;
Darwin)
PLATFORM=OS_MACOSX
COMMON_FLAGS="$MEMCMP_FLAG -DOS_MACOSX"
PLATFORM_SHARED_EXT=dylib
[ -z "$INSTALL_PATH" ] && INSTALL_PATH=`pwd`
PLATFORM_SHARED_LDFLAGS="-dynamiclib -install_name $INSTALL_PATH/"
PORT_FILE=port/port_posix.cc
;;
Linux)
PLATFORM=OS_LINUX
COMMON_FLAGS="$MEMCMP_FLAG -pthread -DOS_LINUX"
PLATFORM_LDFLAGS="-pthread"
PORT_FILE=port/port_posix.cc
;;
SunOS)
PLATFORM=OS_SOLARIS
COMMON_FLAGS="$MEMCMP_FLAG -D_REENTRANT -DOS_SOLARIS"
PLATFORM_LIBS="-lpthread -lrt"
PORT_FILE=port/port_posix.cc
;;
FreeBSD)
PLATFORM=OS_FREEBSD
COMMON_FLAGS="$MEMCMP_FLAG -D_REENTRANT -DOS_FREEBSD"
PLATFORM_LIBS="-lpthread"
PORT_FILE=port/port_posix.cc
;;
NetBSD)
PLATFORM=OS_NETBSD
COMMON_FLAGS="$MEMCMP_FLAG -D_REENTRANT -DOS_NETBSD"
PLATFORM_LIBS="-lpthread -lgcc_s"
PORT_FILE=port/port_posix.cc
;;
OpenBSD)
PLATFORM=OS_OPENBSD
COMMON_FLAGS="$MEMCMP_FLAG -D_REENTRANT -DOS_OPENBSD"
PLATFORM_LDFLAGS="-pthread"
PORT_FILE=port/port_posix.cc
;;
DragonFly)
PLATFORM=OS_DRAGONFLYBSD
COMMON_FLAGS="$MEMCMP_FLAG -D_REENTRANT -DOS_DRAGONFLYBSD"
PLATFORM_LIBS="-lpthread"
PORT_FILE=port/port_posix.cc
;;
OS_ANDROID_CROSSCOMPILE)
PLATFORM=OS_ANDROID
COMMON_FLAGS="$MEMCMP_FLAG -D_REENTRANT -DOS_ANDROID -DLEVELDB_PLATFORM_POSIX"
PLATFORM_LDFLAGS="" # All pthread features are in the Android C library
PORT_FILE=port/port_posix.cc
CROSS_COMPILE=true
;;
HP-UX)
PLATFORM=OS_HPUX
COMMON_FLAGS="$MEMCMP_FLAG -D_REENTRANT -DOS_HPUX"
PLATFORM_LDFLAGS="-pthread"
PORT_FILE=port/port_posix.cc
# man ld: +h internal_name
PLATFORM_SHARED_LDFLAGS="-shared -Wl,+h -Wl,"
;;
IOS)
PLATFORM=IOS
COMMON_FLAGS="$MEMCMP_FLAG -DOS_MACOSX"
[ -z "$INSTALL_PATH" ] && INSTALL_PATH=`pwd`
PORT_FILE=port/port_posix.cc
PLATFORM_SHARED_EXT=
PLATFORM_SHARED_LDFLAGS=
PLATFORM_SHARED_CFLAGS=
PLATFORM_SHARED_VERSIONED=
;;
*)
echo "Unknown platform!" >&2
exit 1
esac
# We want to make a list of all cc files within util, db, table, and helpers
# except for the test and benchmark files. By default, find will output a list
# of all files matching either rule, so we need to append -print to make the
# prune take effect.
DIRS="$PREFIX/db $PREFIX/util $PREFIX/table"
set -f # temporarily disable globbing so that our patterns aren't expanded
PRUNE_TEST="-name *test*.cc -prune"
PRUNE_BENCH="-name *_bench.cc -prune"
PRUNE_TOOL="-name leveldb_main.cc -prune"
PORTABLE_FILES=`find $DIRS $PRUNE_TEST -o $PRUNE_BENCH -o $PRUNE_TOOL -o -name '*.cc' -print | sort | sed "s,^$PREFIX/,," | tr "\n" " "`
set +f # re-enable globbing
# The sources consist of the portable files, plus the platform-specific port
# file.
echo "SOURCES=$PORTABLE_FILES $PORT_FILE" >> $OUTPUT
echo "MEMENV_SOURCES=helpers/memenv/memenv.cc" >> $OUTPUT
if [ "$CROSS_COMPILE" = "true" ]; then
# Cross-compiling; do not try any compilation tests.
true
else
CXXOUTPUT="${TMPDIR}/leveldb_build_detect_platform-cxx.$$"
# If -std=c++0x works, use <atomic> as fallback for when memory barriers
# are not available.
$CXX $CXXFLAGS -std=c++0x -x c++ - -o $CXXOUTPUT 2>/dev/null <<EOF
#include <atomic>
int main() {}
EOF
if [ "$?" = 0 ]; then
COMMON_FLAGS="$COMMON_FLAGS -DLEVELDB_PLATFORM_POSIX -DLEVELDB_ATOMIC_PRESENT"
PLATFORM_CXXFLAGS="-std=c++0x"
else
COMMON_FLAGS="$COMMON_FLAGS -DLEVELDB_PLATFORM_POSIX"
fi
# Test whether Snappy library is installed
# http://code.google.com/p/snappy/
$CXX $CXXFLAGS -x c++ - -o $CXXOUTPUT 2>/dev/null <<EOF
#include <snappy.h>
int main() {}
EOF
if [ "$?" = 0 ]; then
COMMON_FLAGS="$COMMON_FLAGS -DSNAPPY"
PLATFORM_LIBS="$PLATFORM_LIBS -lsnappy"
fi
# Test whether tcmalloc is available
$CXX $CXXFLAGS -x c++ - -o $CXXOUTPUT -ltcmalloc 2>/dev/null <<EOF
int main() {}
EOF
if [ "$?" = 0 ]; then
PLATFORM_LIBS="$PLATFORM_LIBS -ltcmalloc"
fi
rm -f $CXXOUTPUT 2>/dev/null
fi
PLATFORM_CCFLAGS="$PLATFORM_CCFLAGS $COMMON_FLAGS"
PLATFORM_CXXFLAGS="$PLATFORM_CXXFLAGS $COMMON_FLAGS"
echo "CC=$CC" >> $OUTPUT
echo "CXX=$CXX" >> $OUTPUT
echo "PLATFORM=$PLATFORM" >> $OUTPUT
echo "PLATFORM_LDFLAGS=$PLATFORM_LDFLAGS" >> $OUTPUT
echo "PLATFORM_LIBS=$PLATFORM_LIBS" >> $OUTPUT
echo "PLATFORM_CCFLAGS=$PLATFORM_CCFLAGS" >> $OUTPUT
echo "PLATFORM_CXXFLAGS=$PLATFORM_CXXFLAGS" >> $OUTPUT
echo "PLATFORM_SHARED_CFLAGS=$PLATFORM_SHARED_CFLAGS" >> $OUTPUT
echo "PLATFORM_SHARED_EXT=$PLATFORM_SHARED_EXT" >> $OUTPUT
echo "PLATFORM_SHARED_LDFLAGS=$PLATFORM_SHARED_LDFLAGS" >> $OUTPUT
echo "PLATFORM_SHARED_VERSIONED=$PLATFORM_SHARED_VERSIONED" >> $OUTPUT

View File

@ -1,9 +0,0 @@
# Copyright 2019 The LevelDB Authors. All rights reserved.
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file. See the AUTHORS file for names of contributors.
@PACKAGE_INIT@
include("${CMAKE_CURRENT_LIST_DIR}/leveldbTargets.cmake")
check_required_components(leveldb)

View File

@ -2,24 +2,29 @@
// Use of this source code is governed by a BSD-style license that can be // Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file. See the AUTHORS file for names of contributors. // found in the LICENSE file. See the AUTHORS file for names of contributors.
#include "gtest/gtest.h" #include "leveldb/db.h"
#include "db/db_impl.h" #include "db/db_impl.h"
#include "leveldb/cache.h" #include "leveldb/cache.h"
#include "leveldb/db.h" #include "util/testharness.h"
#include "util/testutil.h" #include "util/testutil.h"
namespace leveldb { namespace leveldb {
class AutoCompactTest : public testing::Test { class AutoCompactTest {
public: public:
std::string dbname_;
Cache* tiny_cache_;
Options options_;
DB* db_;
AutoCompactTest() { AutoCompactTest() {
dbname_ = testing::TempDir() + "autocompact_test"; dbname_ = test::TmpDir() + "/autocompact_test";
tiny_cache_ = NewLRUCache(100); tiny_cache_ = NewLRUCache(100);
options_.block_cache = tiny_cache_; options_.block_cache = tiny_cache_;
DestroyDB(dbname_, options_); DestroyDB(dbname_, options_);
options_.create_if_missing = true; options_.create_if_missing = true;
options_.compression = kNoCompression; options_.compression = kNoCompression;
EXPECT_LEVELDB_OK(DB::Open(options_, dbname_, &db_)); ASSERT_OK(DB::Open(options_, dbname_, &db_));
} }
~AutoCompactTest() { ~AutoCompactTest() {
@ -30,7 +35,7 @@ class AutoCompactTest : public testing::Test {
std::string Key(int i) { std::string Key(int i) {
char buf[100]; char buf[100];
std::snprintf(buf, sizeof(buf), "key%06d", i); snprintf(buf, sizeof(buf), "key%06d", i);
return std::string(buf); return std::string(buf);
} }
@ -42,12 +47,6 @@ class AutoCompactTest : public testing::Test {
} }
void DoReads(int n); void DoReads(int n);
private:
std::string dbname_;
Cache* tiny_cache_;
Options options_;
DB* db_;
}; };
static const int kValueSize = 200 * 1024; static const int kValueSize = 200 * 1024;
@ -62,15 +61,15 @@ void AutoCompactTest::DoReads(int n) {
// Fill database // Fill database
for (int i = 0; i < kCount; i++) { for (int i = 0; i < kCount; i++) {
ASSERT_LEVELDB_OK(db_->Put(WriteOptions(), Key(i), value)); ASSERT_OK(db_->Put(WriteOptions(), Key(i), value));
} }
ASSERT_LEVELDB_OK(dbi->TEST_CompactMemTable()); ASSERT_OK(dbi->TEST_CompactMemTable());
// Delete everything // Delete everything
for (int i = 0; i < kCount; i++) { for (int i = 0; i < kCount; i++) {
ASSERT_LEVELDB_OK(db_->Delete(WriteOptions(), Key(i))); ASSERT_OK(db_->Delete(WriteOptions(), Key(i)));
} }
ASSERT_LEVELDB_OK(dbi->TEST_CompactMemTable()); ASSERT_OK(dbi->TEST_CompactMemTable());
// Get initial measurement of the space we will be reading. // Get initial measurement of the space we will be reading.
const int64_t initial_size = Size(Key(0), Key(n)); const int64_t initial_size = Size(Key(0), Key(n));
@ -82,16 +81,17 @@ void AutoCompactTest::DoReads(int n) {
ASSERT_LT(read, 100) << "Taking too long to compact"; ASSERT_LT(read, 100) << "Taking too long to compact";
Iterator* iter = db_->NewIterator(ReadOptions()); Iterator* iter = db_->NewIterator(ReadOptions());
for (iter->SeekToFirst(); for (iter->SeekToFirst();
iter->Valid() && iter->key().ToString() < limit_key; iter->Next()) { iter->Valid() && iter->key().ToString() < limit_key;
iter->Next()) {
// Drop data // Drop data
} }
delete iter; delete iter;
// Wait a little bit to allow any triggered compactions to complete. // Wait a little bit to allow any triggered compactions to complete.
Env::Default()->SleepForMicroseconds(1000000); Env::Default()->SleepForMicroseconds(1000000);
uint64_t size = Size(Key(0), Key(n)); uint64_t size = Size(Key(0), Key(n));
std::fprintf(stderr, "iter %3d => %7.3f MB [other %7.3f MB]\n", read + 1, fprintf(stderr, "iter %3d => %7.3f MB [other %7.3f MB]\n",
size / 1048576.0, Size(Key(n), Key(kCount)) / 1048576.0); read+1, size/1048576.0, Size(Key(n), Key(kCount))/1048576.0);
if (size <= initial_size / 10) { if (size <= initial_size/10) {
break; break;
} }
} }
@ -100,11 +100,19 @@ void AutoCompactTest::DoReads(int n) {
// is pretty much unchanged. // is pretty much unchanged.
const int64_t final_other_size = Size(Key(n), Key(kCount)); const int64_t final_other_size = Size(Key(n), Key(kCount));
ASSERT_LE(final_other_size, initial_other_size + 1048576); ASSERT_LE(final_other_size, initial_other_size + 1048576);
ASSERT_GE(final_other_size, initial_other_size / 5 - 1048576); ASSERT_GE(final_other_size, initial_other_size/5 - 1048576);
} }
TEST_F(AutoCompactTest, ReadAll) { DoReads(kCount); } TEST(AutoCompactTest, ReadAll) {
DoReads(kCount);
}
TEST_F(AutoCompactTest, ReadHalf) { DoReads(kCount / 2); } TEST(AutoCompactTest, ReadHalf) {
DoReads(kCount/2);
}
} // namespace leveldb } // namespace leveldb
int main(int argc, char** argv) {
return leveldb::test::RunAllTests();
}

View File

@ -4,8 +4,8 @@
#include "db/builder.h" #include "db/builder.h"
#include "db/dbformat.h"
#include "db/filename.h" #include "db/filename.h"
#include "db/dbformat.h"
#include "db/table_cache.h" #include "db/table_cache.h"
#include "db/version_edit.h" #include "db/version_edit.h"
#include "leveldb/db.h" #include "leveldb/db.h"
@ -14,8 +14,12 @@
namespace leveldb { namespace leveldb {
Status BuildTable(const std::string& dbname, Env* env, const Options& options, Status BuildTable(const std::string& dbname,
TableCache* table_cache, Iterator* iter, FileMetaData* meta) { Env* env,
const Options& options,
TableCache* table_cache,
Iterator* iter,
FileMetaData* meta) {
Status s; Status s;
meta->file_size = 0; meta->file_size = 0;
iter->SeekToFirst(); iter->SeekToFirst();
@ -30,20 +34,21 @@ Status BuildTable(const std::string& dbname, Env* env, const Options& options,
TableBuilder* builder = new TableBuilder(options, file); TableBuilder* builder = new TableBuilder(options, file);
meta->smallest.DecodeFrom(iter->key()); meta->smallest.DecodeFrom(iter->key());
Slice key;
for (; iter->Valid(); iter->Next()) { for (; iter->Valid(); iter->Next()) {
key = iter->key(); Slice key = iter->key();
builder->Add(key, iter->value());
}
if (!key.empty()) {
meta->largest.DecodeFrom(key); meta->largest.DecodeFrom(key);
builder->Add(key, iter->value());
} }
// Finish and check for builder errors // Finish and check for builder errors
s = builder->Finish();
if (s.ok()) { if (s.ok()) {
meta->file_size = builder->FileSize(); s = builder->Finish();
assert(meta->file_size > 0); if (s.ok()) {
meta->file_size = builder->FileSize();
assert(meta->file_size > 0);
}
} else {
builder->Abandon();
} }
delete builder; delete builder;
@ -55,11 +60,12 @@ Status BuildTable(const std::string& dbname, Env* env, const Options& options,
s = file->Close(); s = file->Close();
} }
delete file; delete file;
file = nullptr; file = NULL;
if (s.ok()) { if (s.ok()) {
// Verify that the table is usable // Verify that the table is usable
Iterator* it = table_cache->NewIterator(ReadOptions(), meta->number, Iterator* it = table_cache->NewIterator(ReadOptions(),
meta->number,
meta->file_size); meta->file_size);
s = it->status(); s = it->status();
delete it; delete it;
@ -74,7 +80,7 @@ Status BuildTable(const std::string& dbname, Env* env, const Options& options,
if (s.ok() && meta->file_size > 0) { if (s.ok() && meta->file_size > 0) {
// Keep it // Keep it
} else { } else {
env->RemoveFile(fname); env->DeleteFile(fname);
} }
return s; return s;
} }

View File

@ -22,8 +22,12 @@ class VersionEdit;
// *meta will be filled with metadata about the generated table. // *meta will be filled with metadata about the generated table.
// If no data is present in *iter, meta->file_size will be set to // If no data is present in *iter, meta->file_size will be set to
// zero, and no Table file will be produced. // zero, and no Table file will be produced.
Status BuildTable(const std::string& dbname, Env* env, const Options& options, extern Status BuildTable(const std::string& dbname,
TableCache* table_cache, Iterator* iter, FileMetaData* meta); Env* env,
const Options& options,
TableCache* table_cache,
Iterator* iter,
FileMetaData* meta);
} // namespace leveldb } // namespace leveldb

402
db/c.cc
View File

@ -4,11 +4,8 @@
#include "leveldb/c.h" #include "leveldb/c.h"
#include <string.h> #include <stdlib.h>
#include <unistd.h>
#include <cstdint>
#include <cstdlib>
#include "leveldb/cache.h" #include "leveldb/cache.h"
#include "leveldb/comparator.h" #include "leveldb/comparator.h"
#include "leveldb/db.h" #include "leveldb/db.h"
@ -46,72 +43,69 @@ using leveldb::WriteOptions;
extern "C" { extern "C" {
struct leveldb_t { struct leveldb_t { DB* rep; };
DB* rep; struct leveldb_iterator_t { Iterator* rep; };
}; struct leveldb_writebatch_t { WriteBatch rep; };
struct leveldb_iterator_t { struct leveldb_snapshot_t { const Snapshot* rep; };
Iterator* rep; struct leveldb_readoptions_t { ReadOptions rep; };
}; struct leveldb_writeoptions_t { WriteOptions rep; };
struct leveldb_writebatch_t { struct leveldb_options_t { Options rep; };
WriteBatch rep; struct leveldb_cache_t { Cache* rep; };
}; struct leveldb_seqfile_t { SequentialFile* rep; };
struct leveldb_snapshot_t { struct leveldb_randomfile_t { RandomAccessFile* rep; };
const Snapshot* rep; struct leveldb_writablefile_t { WritableFile* rep; };
}; struct leveldb_logger_t { Logger* rep; };
struct leveldb_readoptions_t { struct leveldb_filelock_t { FileLock* rep; };
ReadOptions rep;
};
struct leveldb_writeoptions_t {
WriteOptions rep;
};
struct leveldb_options_t {
Options rep;
};
struct leveldb_cache_t {
Cache* rep;
};
struct leveldb_seqfile_t {
SequentialFile* rep;
};
struct leveldb_randomfile_t {
RandomAccessFile* rep;
};
struct leveldb_writablefile_t {
WritableFile* rep;
};
struct leveldb_logger_t {
Logger* rep;
};
struct leveldb_filelock_t {
FileLock* rep;
};
struct leveldb_comparator_t : public Comparator { struct leveldb_comparator_t : public Comparator {
~leveldb_comparator_t() override { (*destructor_)(state_); } void* state_;
void (*destructor_)(void*);
int (*compare_)(
void*,
const char* a, size_t alen,
const char* b, size_t blen);
const char* (*name_)(void*);
int Compare(const Slice& a, const Slice& b) const override { virtual ~leveldb_comparator_t() {
(*destructor_)(state_);
}
virtual int Compare(const Slice& a, const Slice& b) const {
return (*compare_)(state_, a.data(), a.size(), b.data(), b.size()); return (*compare_)(state_, a.data(), a.size(), b.data(), b.size());
} }
const char* Name() const override { return (*name_)(state_); } virtual const char* Name() const {
return (*name_)(state_);
}
// No-ops since the C binding does not support key shortening methods. // No-ops since the C binding does not support key shortening methods.
void FindShortestSeparator(std::string*, const Slice&) const override {} virtual void FindShortestSeparator(std::string*, const Slice&) const { }
void FindShortSuccessor(std::string* key) const override {} virtual void FindShortSuccessor(std::string* key) const { }
void* state_;
void (*destructor_)(void*);
int (*compare_)(void*, const char* a, size_t alen, const char* b,
size_t blen);
const char* (*name_)(void*);
}; };
struct leveldb_filterpolicy_t : public FilterPolicy { struct leveldb_filterpolicy_t : public FilterPolicy {
~leveldb_filterpolicy_t() override { (*destructor_)(state_); } void* state_;
void (*destructor_)(void*);
const char* (*name_)(void*);
char* (*create_)(
void*,
const char* const* key_array, const size_t* key_length_array,
int num_keys,
size_t* filter_length);
unsigned char (*key_match_)(
void*,
const char* key, size_t length,
const char* filter, size_t filter_length);
const char* Name() const override { return (*name_)(state_); } virtual ~leveldb_filterpolicy_t() {
(*destructor_)(state_);
}
void CreateFilter(const Slice* keys, int n, std::string* dst) const override { virtual const char* Name() const {
return (*name_)(state_);
}
virtual void CreateFilter(const Slice* keys, int n, std::string* dst) const {
std::vector<const char*> key_pointers(n); std::vector<const char*> key_pointers(n);
std::vector<size_t> key_sizes(n); std::vector<size_t> key_sizes(n);
for (int i = 0; i < n; i++) { for (int i = 0; i < n; i++) {
@ -121,22 +115,13 @@ struct leveldb_filterpolicy_t : public FilterPolicy {
size_t len; size_t len;
char* filter = (*create_)(state_, &key_pointers[0], &key_sizes[0], n, &len); char* filter = (*create_)(state_, &key_pointers[0], &key_sizes[0], n, &len);
dst->append(filter, len); dst->append(filter, len);
std::free(filter); free(filter);
} }
bool KeyMayMatch(const Slice& key, const Slice& filter) const override { virtual bool KeyMayMatch(const Slice& key, const Slice& filter) const {
return (*key_match_)(state_, key.data(), key.size(), filter.data(), return (*key_match_)(state_, key.data(), key.size(),
filter.size()); filter.data(), filter.size());
} }
void* state_;
void (*destructor_)(void*);
const char* (*name_)(void*);
char* (*create_)(void*, const char* const* key_array,
const size_t* key_length_array, int num_keys,
size_t* filter_length);
uint8_t (*key_match_)(void*, const char* key, size_t length,
const char* filter, size_t filter_length);
}; };
struct leveldb_env_t { struct leveldb_env_t {
@ -145,31 +130,32 @@ struct leveldb_env_t {
}; };
static bool SaveError(char** errptr, const Status& s) { static bool SaveError(char** errptr, const Status& s) {
assert(errptr != nullptr); assert(errptr != NULL);
if (s.ok()) { if (s.ok()) {
return false; return false;
} else if (*errptr == nullptr) { } else if (*errptr == NULL) {
*errptr = strdup(s.ToString().c_str()); *errptr = strdup(s.ToString().c_str());
} else { } else {
// TODO(sanjay): Merge with existing error? // TODO(sanjay): Merge with existing error?
std::free(*errptr); free(*errptr);
*errptr = strdup(s.ToString().c_str()); *errptr = strdup(s.ToString().c_str());
} }
return true; return true;
} }
static char* CopyString(const std::string& str) { static char* CopyString(const std::string& str) {
char* result = char* result = reinterpret_cast<char*>(malloc(sizeof(char) * str.size()));
reinterpret_cast<char*>(std::malloc(sizeof(char) * str.size())); memcpy(result, str.data(), sizeof(char) * str.size());
std::memcpy(result, str.data(), sizeof(char) * str.size());
return result; return result;
} }
leveldb_t* leveldb_open(const leveldb_options_t* options, const char* name, leveldb_t* leveldb_open(
char** errptr) { const leveldb_options_t* options,
const char* name,
char** errptr) {
DB* db; DB* db;
if (SaveError(errptr, DB::Open(options->rep, std::string(name), &db))) { if (SaveError(errptr, DB::Open(options->rep, std::string(name), &db))) {
return nullptr; return NULL;
} }
leveldb_t* result = new leveldb_t; leveldb_t* result = new leveldb_t;
result->rep = db; result->rep = db;
@ -181,27 +167,40 @@ void leveldb_close(leveldb_t* db) {
delete db; delete db;
} }
void leveldb_put(leveldb_t* db, const leveldb_writeoptions_t* options, void leveldb_put(
const char* key, size_t keylen, const char* val, size_t vallen, leveldb_t* db,
char** errptr) { const leveldb_writeoptions_t* options,
const char* key, size_t keylen,
const char* val, size_t vallen,
char** errptr) {
SaveError(errptr, SaveError(errptr,
db->rep->Put(options->rep, Slice(key, keylen), Slice(val, vallen))); db->rep->Put(options->rep, Slice(key, keylen), Slice(val, vallen)));
} }
void leveldb_delete(leveldb_t* db, const leveldb_writeoptions_t* options, void leveldb_delete(
const char* key, size_t keylen, char** errptr) { leveldb_t* db,
const leveldb_writeoptions_t* options,
const char* key, size_t keylen,
char** errptr) {
SaveError(errptr, db->rep->Delete(options->rep, Slice(key, keylen))); SaveError(errptr, db->rep->Delete(options->rep, Slice(key, keylen)));
} }
void leveldb_write(leveldb_t* db, const leveldb_writeoptions_t* options,
leveldb_writebatch_t* batch, char** errptr) { void leveldb_write(
leveldb_t* db,
const leveldb_writeoptions_t* options,
leveldb_writebatch_t* batch,
char** errptr) {
SaveError(errptr, db->rep->Write(options->rep, &batch->rep)); SaveError(errptr, db->rep->Write(options->rep, &batch->rep));
} }
char* leveldb_get(leveldb_t* db, const leveldb_readoptions_t* options, char* leveldb_get(
const char* key, size_t keylen, size_t* vallen, leveldb_t* db,
char** errptr) { const leveldb_readoptions_t* options,
char* result = nullptr; const char* key, size_t keylen,
size_t* vallen,
char** errptr) {
char* result = NULL;
std::string tmp; std::string tmp;
Status s = db->rep->Get(options->rep, Slice(key, keylen), &tmp); Status s = db->rep->Get(options->rep, Slice(key, keylen), &tmp);
if (s.ok()) { if (s.ok()) {
@ -217,40 +216,45 @@ char* leveldb_get(leveldb_t* db, const leveldb_readoptions_t* options,
} }
leveldb_iterator_t* leveldb_create_iterator( leveldb_iterator_t* leveldb_create_iterator(
leveldb_t* db, const leveldb_readoptions_t* options) { leveldb_t* db,
const leveldb_readoptions_t* options) {
leveldb_iterator_t* result = new leveldb_iterator_t; leveldb_iterator_t* result = new leveldb_iterator_t;
result->rep = db->rep->NewIterator(options->rep); result->rep = db->rep->NewIterator(options->rep);
return result; return result;
} }
const leveldb_snapshot_t* leveldb_create_snapshot(leveldb_t* db) { const leveldb_snapshot_t* leveldb_create_snapshot(
leveldb_t* db) {
leveldb_snapshot_t* result = new leveldb_snapshot_t; leveldb_snapshot_t* result = new leveldb_snapshot_t;
result->rep = db->rep->GetSnapshot(); result->rep = db->rep->GetSnapshot();
return result; return result;
} }
void leveldb_release_snapshot(leveldb_t* db, void leveldb_release_snapshot(
const leveldb_snapshot_t* snapshot) { leveldb_t* db,
const leveldb_snapshot_t* snapshot) {
db->rep->ReleaseSnapshot(snapshot->rep); db->rep->ReleaseSnapshot(snapshot->rep);
delete snapshot; delete snapshot;
} }
char* leveldb_property_value(leveldb_t* db, const char* propname) { char* leveldb_property_value(
leveldb_t* db,
const char* propname) {
std::string tmp; std::string tmp;
if (db->rep->GetProperty(Slice(propname), &tmp)) { if (db->rep->GetProperty(Slice(propname), &tmp)) {
// We use strdup() since we expect human readable output. // We use strdup() since we expect human readable output.
return strdup(tmp.c_str()); return strdup(tmp.c_str());
} else { } else {
return nullptr; return NULL;
} }
} }
void leveldb_approximate_sizes(leveldb_t* db, int num_ranges, void leveldb_approximate_sizes(
const char* const* range_start_key, leveldb_t* db,
const size_t* range_start_key_len, int num_ranges,
const char* const* range_limit_key, const char* const* range_start_key, const size_t* range_start_key_len,
const size_t* range_limit_key_len, const char* const* range_limit_key, const size_t* range_limit_key_len,
uint64_t* sizes) { uint64_t* sizes) {
Range* ranges = new Range[num_ranges]; Range* ranges = new Range[num_ranges];
for (int i = 0; i < num_ranges; i++) { for (int i = 0; i < num_ranges; i++) {
ranges[i].start = Slice(range_start_key[i], range_start_key_len[i]); ranges[i].start = Slice(range_start_key[i], range_start_key_len[i]);
@ -260,23 +264,28 @@ void leveldb_approximate_sizes(leveldb_t* db, int num_ranges,
delete[] ranges; delete[] ranges;
} }
void leveldb_compact_range(leveldb_t* db, const char* start_key, void leveldb_compact_range(
size_t start_key_len, const char* limit_key, leveldb_t* db,
size_t limit_key_len) { const char* start_key, size_t start_key_len,
const char* limit_key, size_t limit_key_len) {
Slice a, b; Slice a, b;
db->rep->CompactRange( db->rep->CompactRange(
// Pass null Slice if corresponding "const char*" is null // Pass NULL Slice if corresponding "const char*" is NULL
(start_key ? (a = Slice(start_key, start_key_len), &a) : nullptr), (start_key ? (a = Slice(start_key, start_key_len), &a) : NULL),
(limit_key ? (b = Slice(limit_key, limit_key_len), &b) : nullptr)); (limit_key ? (b = Slice(limit_key, limit_key_len), &b) : NULL));
} }
void leveldb_destroy_db(const leveldb_options_t* options, const char* name, void leveldb_destroy_db(
char** errptr) { const leveldb_options_t* options,
const char* name,
char** errptr) {
SaveError(errptr, DestroyDB(name, options->rep)); SaveError(errptr, DestroyDB(name, options->rep));
} }
void leveldb_repair_db(const leveldb_options_t* options, const char* name, void leveldb_repair_db(
char** errptr) { const leveldb_options_t* options,
const char* name,
char** errptr) {
SaveError(errptr, RepairDB(name, options->rep)); SaveError(errptr, RepairDB(name, options->rep));
} }
@ -285,7 +294,7 @@ void leveldb_iter_destroy(leveldb_iterator_t* iter) {
delete iter; delete iter;
} }
uint8_t leveldb_iter_valid(const leveldb_iterator_t* iter) { unsigned char leveldb_iter_valid(const leveldb_iterator_t* iter) {
return iter->rep->Valid(); return iter->rep->Valid();
} }
@ -301,9 +310,13 @@ void leveldb_iter_seek(leveldb_iterator_t* iter, const char* k, size_t klen) {
iter->rep->Seek(Slice(k, klen)); iter->rep->Seek(Slice(k, klen));
} }
void leveldb_iter_next(leveldb_iterator_t* iter) { iter->rep->Next(); } void leveldb_iter_next(leveldb_iterator_t* iter) {
iter->rep->Next();
}
void leveldb_iter_prev(leveldb_iterator_t* iter) { iter->rep->Prev(); } void leveldb_iter_prev(leveldb_iterator_t* iter) {
iter->rep->Prev();
}
const char* leveldb_iter_key(const leveldb_iterator_t* iter, size_t* klen) { const char* leveldb_iter_key(const leveldb_iterator_t* iter, size_t* klen) {
Slice s = iter->rep->key(); Slice s = iter->rep->key();
@ -325,34 +338,41 @@ leveldb_writebatch_t* leveldb_writebatch_create() {
return new leveldb_writebatch_t; return new leveldb_writebatch_t;
} }
void leveldb_writebatch_destroy(leveldb_writebatch_t* b) { delete b; } void leveldb_writebatch_destroy(leveldb_writebatch_t* b) {
delete b;
}
void leveldb_writebatch_clear(leveldb_writebatch_t* b) { b->rep.Clear(); } void leveldb_writebatch_clear(leveldb_writebatch_t* b) {
b->rep.Clear();
}
void leveldb_writebatch_put(leveldb_writebatch_t* b, const char* key, void leveldb_writebatch_put(
size_t klen, const char* val, size_t vlen) { leveldb_writebatch_t* b,
const char* key, size_t klen,
const char* val, size_t vlen) {
b->rep.Put(Slice(key, klen), Slice(val, vlen)); b->rep.Put(Slice(key, klen), Slice(val, vlen));
} }
void leveldb_writebatch_delete(leveldb_writebatch_t* b, const char* key, void leveldb_writebatch_delete(
size_t klen) { leveldb_writebatch_t* b,
const char* key, size_t klen) {
b->rep.Delete(Slice(key, klen)); b->rep.Delete(Slice(key, klen));
} }
void leveldb_writebatch_iterate(const leveldb_writebatch_t* b, void* state, void leveldb_writebatch_iterate(
void (*put)(void*, const char* k, size_t klen, leveldb_writebatch_t* b,
const char* v, size_t vlen), void* state,
void (*deleted)(void*, const char* k, void (*put)(void*, const char* k, size_t klen, const char* v, size_t vlen),
size_t klen)) { void (*deleted)(void*, const char* k, size_t klen)) {
class H : public WriteBatch::Handler { class H : public WriteBatch::Handler {
public: public:
void* state_; void* state_;
void (*put_)(void*, const char* k, size_t klen, const char* v, size_t vlen); void (*put_)(void*, const char* k, size_t klen, const char* v, size_t vlen);
void (*deleted_)(void*, const char* k, size_t klen); void (*deleted_)(void*, const char* k, size_t klen);
void Put(const Slice& key, const Slice& value) override { virtual void Put(const Slice& key, const Slice& value) {
(*put_)(state_, key.data(), key.size(), value.data(), value.size()); (*put_)(state_, key.data(), key.size(), value.data(), value.size());
} }
void Delete(const Slice& key) override { virtual void Delete(const Slice& key) {
(*deleted_)(state_, key.data(), key.size()); (*deleted_)(state_, key.data(), key.size());
} }
}; };
@ -363,43 +383,47 @@ void leveldb_writebatch_iterate(const leveldb_writebatch_t* b, void* state,
b->rep.Iterate(&handler); b->rep.Iterate(&handler);
} }
void leveldb_writebatch_append(leveldb_writebatch_t* destination, leveldb_options_t* leveldb_options_create() {
const leveldb_writebatch_t* source) { return new leveldb_options_t;
destination->rep.Append(source->rep);
} }
leveldb_options_t* leveldb_options_create() { return new leveldb_options_t; } void leveldb_options_destroy(leveldb_options_t* options) {
delete options;
}
void leveldb_options_destroy(leveldb_options_t* options) { delete options; } void leveldb_options_set_comparator(
leveldb_options_t* opt,
void leveldb_options_set_comparator(leveldb_options_t* opt, leveldb_comparator_t* cmp) {
leveldb_comparator_t* cmp) {
opt->rep.comparator = cmp; opt->rep.comparator = cmp;
} }
void leveldb_options_set_filter_policy(leveldb_options_t* opt, void leveldb_options_set_filter_policy(
leveldb_filterpolicy_t* policy) { leveldb_options_t* opt,
leveldb_filterpolicy_t* policy) {
opt->rep.filter_policy = policy; opt->rep.filter_policy = policy;
} }
void leveldb_options_set_create_if_missing(leveldb_options_t* opt, uint8_t v) { void leveldb_options_set_create_if_missing(
leveldb_options_t* opt, unsigned char v) {
opt->rep.create_if_missing = v; opt->rep.create_if_missing = v;
} }
void leveldb_options_set_error_if_exists(leveldb_options_t* opt, uint8_t v) { void leveldb_options_set_error_if_exists(
leveldb_options_t* opt, unsigned char v) {
opt->rep.error_if_exists = v; opt->rep.error_if_exists = v;
} }
void leveldb_options_set_paranoid_checks(leveldb_options_t* opt, uint8_t v) { void leveldb_options_set_paranoid_checks(
leveldb_options_t* opt, unsigned char v) {
opt->rep.paranoid_checks = v; opt->rep.paranoid_checks = v;
} }
void leveldb_options_set_env(leveldb_options_t* opt, leveldb_env_t* env) { void leveldb_options_set_env(leveldb_options_t* opt, leveldb_env_t* env) {
opt->rep.env = (env ? env->rep : nullptr); opt->rep.env = (env ? env->rep : NULL);
} }
void leveldb_options_set_info_log(leveldb_options_t* opt, leveldb_logger_t* l) { void leveldb_options_set_info_log(leveldb_options_t* opt, leveldb_logger_t* l) {
opt->rep.info_log = (l ? l->rep : nullptr); opt->rep.info_log = (l ? l->rep : NULL);
} }
void leveldb_options_set_write_buffer_size(leveldb_options_t* opt, size_t s) { void leveldb_options_set_write_buffer_size(leveldb_options_t* opt, size_t s) {
@ -422,18 +446,17 @@ void leveldb_options_set_block_restart_interval(leveldb_options_t* opt, int n) {
opt->rep.block_restart_interval = n; opt->rep.block_restart_interval = n;
} }
void leveldb_options_set_max_file_size(leveldb_options_t* opt, size_t s) {
opt->rep.max_file_size = s;
}
void leveldb_options_set_compression(leveldb_options_t* opt, int t) { void leveldb_options_set_compression(leveldb_options_t* opt, int t) {
opt->rep.compression = static_cast<CompressionType>(t); opt->rep.compression = static_cast<CompressionType>(t);
} }
leveldb_comparator_t* leveldb_comparator_create( leveldb_comparator_t* leveldb_comparator_create(
void* state, void (*destructor)(void*), void* state,
int (*compare)(void*, const char* a, size_t alen, const char* b, void (*destructor)(void*),
size_t blen), int (*compare)(
void*,
const char* a, size_t alen,
const char* b, size_t blen),
const char* (*name)(void*)) { const char* (*name)(void*)) {
leveldb_comparator_t* result = new leveldb_comparator_t; leveldb_comparator_t* result = new leveldb_comparator_t;
result->state_ = state; result->state_ = state;
@ -443,15 +466,22 @@ leveldb_comparator_t* leveldb_comparator_create(
return result; return result;
} }
void leveldb_comparator_destroy(leveldb_comparator_t* cmp) { delete cmp; } void leveldb_comparator_destroy(leveldb_comparator_t* cmp) {
delete cmp;
}
leveldb_filterpolicy_t* leveldb_filterpolicy_create( leveldb_filterpolicy_t* leveldb_filterpolicy_create(
void* state, void (*destructor)(void*), void* state,
char* (*create_filter)(void*, const char* const* key_array, void (*destructor)(void*),
const size_t* key_length_array, int num_keys, char* (*create_filter)(
size_t* filter_length), void*,
uint8_t (*key_may_match)(void*, const char* key, size_t length, const char* const* key_array, const size_t* key_length_array,
const char* filter, size_t filter_length), int num_keys,
size_t* filter_length),
unsigned char (*key_may_match)(
void*,
const char* key, size_t length,
const char* filter, size_t filter_length),
const char* (*name)(void*)) { const char* (*name)(void*)) {
leveldb_filterpolicy_t* result = new leveldb_filterpolicy_t; leveldb_filterpolicy_t* result = new leveldb_filterpolicy_t;
result->state_ = state; result->state_ = state;
@ -471,8 +501,7 @@ leveldb_filterpolicy_t* leveldb_filterpolicy_create_bloom(int bits_per_key) {
// they delegate to a NewBloomFilterPolicy() instead of user // they delegate to a NewBloomFilterPolicy() instead of user
// supplied C functions. // supplied C functions.
struct Wrapper : public leveldb_filterpolicy_t { struct Wrapper : public leveldb_filterpolicy_t {
static void DoNothing(void*) {} const FilterPolicy* rep_;
~Wrapper() { delete rep_; } ~Wrapper() { delete rep_; }
const char* Name() const { return rep_->Name(); } const char* Name() const { return rep_->Name(); }
void CreateFilter(const Slice* keys, int n, std::string* dst) const { void CreateFilter(const Slice* keys, int n, std::string* dst) const {
@ -481,12 +510,11 @@ leveldb_filterpolicy_t* leveldb_filterpolicy_create_bloom(int bits_per_key) {
bool KeyMayMatch(const Slice& key, const Slice& filter) const { bool KeyMayMatch(const Slice& key, const Slice& filter) const {
return rep_->KeyMayMatch(key, filter); return rep_->KeyMayMatch(key, filter);
} }
static void DoNothing(void*) { }
const FilterPolicy* rep_;
}; };
Wrapper* wrapper = new Wrapper; Wrapper* wrapper = new Wrapper;
wrapper->rep_ = NewBloomFilterPolicy(bits_per_key); wrapper->rep_ = NewBloomFilterPolicy(bits_per_key);
wrapper->state_ = nullptr; wrapper->state_ = NULL;
wrapper->destructor_ = &Wrapper::DoNothing; wrapper->destructor_ = &Wrapper::DoNothing;
return wrapper; return wrapper;
} }
@ -495,29 +523,37 @@ leveldb_readoptions_t* leveldb_readoptions_create() {
return new leveldb_readoptions_t; return new leveldb_readoptions_t;
} }
void leveldb_readoptions_destroy(leveldb_readoptions_t* opt) { delete opt; } void leveldb_readoptions_destroy(leveldb_readoptions_t* opt) {
delete opt;
}
void leveldb_readoptions_set_verify_checksums(leveldb_readoptions_t* opt, void leveldb_readoptions_set_verify_checksums(
uint8_t v) { leveldb_readoptions_t* opt,
unsigned char v) {
opt->rep.verify_checksums = v; opt->rep.verify_checksums = v;
} }
void leveldb_readoptions_set_fill_cache(leveldb_readoptions_t* opt, uint8_t v) { void leveldb_readoptions_set_fill_cache(
leveldb_readoptions_t* opt, unsigned char v) {
opt->rep.fill_cache = v; opt->rep.fill_cache = v;
} }
void leveldb_readoptions_set_snapshot(leveldb_readoptions_t* opt, void leveldb_readoptions_set_snapshot(
const leveldb_snapshot_t* snap) { leveldb_readoptions_t* opt,
opt->rep.snapshot = (snap ? snap->rep : nullptr); const leveldb_snapshot_t* snap) {
opt->rep.snapshot = (snap ? snap->rep : NULL);
} }
leveldb_writeoptions_t* leveldb_writeoptions_create() { leveldb_writeoptions_t* leveldb_writeoptions_create() {
return new leveldb_writeoptions_t; return new leveldb_writeoptions_t;
} }
void leveldb_writeoptions_destroy(leveldb_writeoptions_t* opt) { delete opt; } void leveldb_writeoptions_destroy(leveldb_writeoptions_t* opt) {
delete opt;
}
void leveldb_writeoptions_set_sync(leveldb_writeoptions_t* opt, uint8_t v) { void leveldb_writeoptions_set_sync(
leveldb_writeoptions_t* opt, unsigned char v) {
opt->rep.sync = v; opt->rep.sync = v;
} }
@ -544,22 +580,16 @@ void leveldb_env_destroy(leveldb_env_t* env) {
delete env; delete env;
} }
char* leveldb_env_get_test_directory(leveldb_env_t* env) { void leveldb_free(void* ptr) {
std::string result; free(ptr);
if (!env->rep->GetTestDirectory(&result).ok()) {
return nullptr;
}
char* buffer = static_cast<char*>(std::malloc(result.size() + 1));
std::memcpy(buffer, result.data(), result.size());
buffer[result.size()] = '\0';
return buffer;
} }
void leveldb_free(void* ptr) { std::free(ptr); } int leveldb_major_version() {
return kMajorVersion;
}
int leveldb_major_version() { return kMajorVersion; } int leveldb_minor_version() {
return kMinorVersion;
int leveldb_minor_version() { return kMinorVersion; } }
} // end extern "C" } // end extern "C"

View File

@ -8,14 +8,24 @@
#include <stdio.h> #include <stdio.h>
#include <stdlib.h> #include <stdlib.h>
#include <string.h> #include <string.h>
#include <sys/types.h>
#include <unistd.h>
const char* phase = ""; const char* phase = "";
static char dbname[200];
static void StartPhase(const char* name) { static void StartPhase(const char* name) {
fprintf(stderr, "=== Test %s\n", name); fprintf(stderr, "=== Test %s\n", name);
phase = name; phase = name;
} }
static const char* GetTempDir(void) {
const char* ret = getenv("TEST_TMPDIR");
if (ret == NULL || ret[0] == '\0')
ret = "/tmp";
return ret;
}
#define CheckNoError(err) \ #define CheckNoError(err) \
if ((err) != NULL) { \ if ((err) != NULL) { \
fprintf(stderr, "%s:%d: %s: %s\n", __FILE__, __LINE__, phase, (err)); \ fprintf(stderr, "%s:%d: %s: %s\n", __FILE__, __LINE__, phase, (err)); \
@ -120,7 +130,7 @@ static const char* CmpName(void* arg) {
} }
// Custom filter policy // Custom filter policy
static uint8_t fake_filter_result = 1; static unsigned char fake_filter_result = 1;
static void FilterDestroy(void* arg) { } static void FilterDestroy(void* arg) { }
static const char* FilterName(void* arg) { static const char* FilterName(void* arg) {
return "TestFilter"; return "TestFilter";
@ -135,8 +145,10 @@ static char* FilterCreate(
memcpy(result, "fake", 4); memcpy(result, "fake", 4);
return result; return result;
} }
uint8_t FilterKeyMatch(void* arg, const char* key, size_t length, unsigned char FilterKeyMatch(
const char* filter, size_t filter_length) { void* arg,
const char* key, size_t length,
const char* filter, size_t filter_length) {
CheckCondition(filter_length == 4); CheckCondition(filter_length == 4);
CheckCondition(memcmp(filter, "fake", 4) == 0); CheckCondition(memcmp(filter, "fake", 4) == 0);
return fake_filter_result; return fake_filter_result;
@ -150,19 +162,21 @@ int main(int argc, char** argv) {
leveldb_options_t* options; leveldb_options_t* options;
leveldb_readoptions_t* roptions; leveldb_readoptions_t* roptions;
leveldb_writeoptions_t* woptions; leveldb_writeoptions_t* woptions;
char* dbname;
char* err = NULL; char* err = NULL;
int run = -1; int run = -1;
CheckCondition(leveldb_major_version() >= 1); CheckCondition(leveldb_major_version() >= 1);
CheckCondition(leveldb_minor_version() >= 1); CheckCondition(leveldb_minor_version() >= 1);
snprintf(dbname, sizeof(dbname),
"%s/leveldb_c_test-%d",
GetTempDir(),
((int) geteuid()));
StartPhase("create_objects"); StartPhase("create_objects");
cmp = leveldb_comparator_create(NULL, CmpDestroy, CmpCompare, CmpName); cmp = leveldb_comparator_create(NULL, CmpDestroy, CmpCompare, CmpName);
env = leveldb_create_default_env(); env = leveldb_create_default_env();
cache = leveldb_cache_create_lru(100000); cache = leveldb_cache_create_lru(100000);
dbname = leveldb_env_get_test_directory(env);
CheckCondition(dbname != NULL);
options = leveldb_options_create(); options = leveldb_options_create();
leveldb_options_set_comparator(options, cmp); leveldb_options_set_comparator(options, cmp);
@ -175,7 +189,6 @@ int main(int argc, char** argv) {
leveldb_options_set_max_open_files(options, 10); leveldb_options_set_max_open_files(options, 10);
leveldb_options_set_block_size(options, 1024); leveldb_options_set_block_size(options, 1024);
leveldb_options_set_block_restart_interval(options, 8); leveldb_options_set_block_restart_interval(options, 8);
leveldb_options_set_max_file_size(options, 3 << 20);
leveldb_options_set_compression(options, leveldb_no_compression); leveldb_options_set_compression(options, leveldb_no_compression);
roptions = leveldb_readoptions_create(); roptions = leveldb_readoptions_create();
@ -226,18 +239,12 @@ int main(int argc, char** argv) {
leveldb_writebatch_clear(wb); leveldb_writebatch_clear(wb);
leveldb_writebatch_put(wb, "bar", 3, "b", 1); leveldb_writebatch_put(wb, "bar", 3, "b", 1);
leveldb_writebatch_put(wb, "box", 3, "c", 1); leveldb_writebatch_put(wb, "box", 3, "c", 1);
leveldb_writebatch_delete(wb, "bar", 3);
leveldb_writebatch_t* wb2 = leveldb_writebatch_create();
leveldb_writebatch_delete(wb2, "bar", 3);
leveldb_writebatch_append(wb, wb2);
leveldb_writebatch_destroy(wb2);
leveldb_write(db, woptions, wb, &err); leveldb_write(db, woptions, wb, &err);
CheckNoError(err); CheckNoError(err);
CheckGet(db, roptions, "foo", "hello"); CheckGet(db, roptions, "foo", "hello");
CheckGet(db, roptions, "bar", NULL); CheckGet(db, roptions, "bar", NULL);
CheckGet(db, roptions, "box", "c"); CheckGet(db, roptions, "box", "c");
int pos = 0; int pos = 0;
leveldb_writebatch_iterate(wb, &pos, CheckPut, CheckDel); leveldb_writebatch_iterate(wb, &pos, CheckPut, CheckDel);
CheckCondition(pos == 3); CheckCondition(pos == 3);
@ -374,7 +381,6 @@ int main(int argc, char** argv) {
leveldb_options_destroy(options); leveldb_options_destroy(options);
leveldb_readoptions_destroy(roptions); leveldb_readoptions_destroy(roptions);
leveldb_writeoptions_destroy(woptions); leveldb_writeoptions_destroy(woptions);
leveldb_free(dbname);
leveldb_cache_destroy(cache); leveldb_cache_destroy(cache);
leveldb_comparator_destroy(cmp); leveldb_comparator_destroy(cmp);
leveldb_env_destroy(env); leveldb_env_destroy(env);

View File

@ -2,63 +2,76 @@
// Use of this source code is governed by a BSD-style license that can be // Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file. See the AUTHORS file for names of contributors. // found in the LICENSE file. See the AUTHORS file for names of contributors.
#include <sys/types.h> #include "leveldb/db.h"
#include "gtest/gtest.h" #include <errno.h>
#include <fcntl.h>
#include <sys/stat.h>
#include <sys/types.h>
#include "leveldb/cache.h"
#include "leveldb/env.h"
#include "leveldb/table.h"
#include "leveldb/write_batch.h"
#include "db/db_impl.h" #include "db/db_impl.h"
#include "db/filename.h" #include "db/filename.h"
#include "db/log_format.h" #include "db/log_format.h"
#include "db/version_set.h" #include "db/version_set.h"
#include "leveldb/cache.h"
#include "leveldb/db.h"
#include "leveldb/table.h"
#include "leveldb/write_batch.h"
#include "util/logging.h" #include "util/logging.h"
#include "util/testharness.h"
#include "util/testutil.h" #include "util/testutil.h"
namespace leveldb { namespace leveldb {
static const int kValueSize = 1000; static const int kValueSize = 1000;
class CorruptionTest : public testing::Test { class CorruptionTest {
public: public:
CorruptionTest() test::ErrorEnv env_;
: db_(nullptr), std::string dbname_;
dbname_("/memenv/corruption_test"), Cache* tiny_cache_;
tiny_cache_(NewLRUCache(100)) { Options options_;
DB* db_;
CorruptionTest() {
tiny_cache_ = NewLRUCache(100);
options_.env = &env_; options_.env = &env_;
options_.block_cache = tiny_cache_; options_.block_cache = tiny_cache_;
dbname_ = test::TmpDir() + "/corruption_test";
DestroyDB(dbname_, options_); DestroyDB(dbname_, options_);
db_ = NULL;
options_.create_if_missing = true; options_.create_if_missing = true;
Reopen(); Reopen();
options_.create_if_missing = false; options_.create_if_missing = false;
} }
~CorruptionTest() { ~CorruptionTest() {
delete db_; delete db_;
delete tiny_cache_; DestroyDB(dbname_, Options());
delete tiny_cache_;
} }
Status TryReopen() { Status TryReopen() {
delete db_; delete db_;
db_ = nullptr; db_ = NULL;
return DB::Open(options_, dbname_, &db_); return DB::Open(options_, dbname_, &db_);
} }
void Reopen() { ASSERT_LEVELDB_OK(TryReopen()); } void Reopen() {
ASSERT_OK(TryReopen());
}
void RepairDB() { void RepairDB() {
delete db_; delete db_;
db_ = nullptr; db_ = NULL;
ASSERT_LEVELDB_OK(::leveldb::RepairDB(dbname_, options_)); ASSERT_OK(::leveldb::RepairDB(dbname_, options_));
} }
void Build(int n) { void Build(int n) {
std::string key_space, value_space; std::string key_space, value_space;
WriteBatch batch; WriteBatch batch;
for (int i = 0; i < n; i++) { for (int i = 0; i < n; i++) {
// if ((i % 100) == 0) std::fprintf(stderr, "@ %d of %d\n", i, n); //if ((i % 100) == 0) fprintf(stderr, "@ %d of %d\n", i, n);
Slice key = Key(i, &key_space); Slice key = Key(i, &key_space);
batch.Clear(); batch.Clear();
batch.Put(key, Value(i, &value_space)); batch.Put(key, Value(i, &value_space));
@ -68,7 +81,7 @@ class CorruptionTest : public testing::Test {
if (i == n - 1) { if (i == n - 1) {
options.sync = true; options.sync = true;
} }
ASSERT_LEVELDB_OK(db_->Write(options, &batch)); ASSERT_OK(db_->Write(options, &batch));
} }
} }
@ -87,7 +100,8 @@ class CorruptionTest : public testing::Test {
// Ignore boundary keys. // Ignore boundary keys.
continue; continue;
} }
if (!ConsumeDecimalNumber(&in, &key) || !in.empty() || if (!ConsumeDecimalNumber(&in, &key) ||
!in.empty() ||
key < next_expected) { key < next_expected) {
bad_keys++; bad_keys++;
continue; continue;
@ -102,10 +116,9 @@ class CorruptionTest : public testing::Test {
} }
delete iter; delete iter;
std::fprintf( fprintf(stderr,
stderr, "expected=%d..%d; got=%d; bad_keys=%d; bad_values=%d; missed=%d\n",
"expected=%d..%d; got=%d; bad_keys=%d; bad_values=%d; missed=%d\n", min_expected, max_expected, correct, bad_keys, bad_values, missed);
min_expected, max_expected, correct, bad_keys, bad_values, missed);
ASSERT_LE(min_expected, correct); ASSERT_LE(min_expected, correct);
ASSERT_GE(max_expected, correct); ASSERT_GE(max_expected, correct);
} }
@ -113,13 +126,14 @@ class CorruptionTest : public testing::Test {
void Corrupt(FileType filetype, int offset, int bytes_to_corrupt) { void Corrupt(FileType filetype, int offset, int bytes_to_corrupt) {
// Pick file to corrupt // Pick file to corrupt
std::vector<std::string> filenames; std::vector<std::string> filenames;
ASSERT_LEVELDB_OK(env_.target()->GetChildren(dbname_, &filenames)); ASSERT_OK(env_.GetChildren(dbname_, &filenames));
uint64_t number; uint64_t number;
FileType type; FileType type;
std::string fname; std::string fname;
int picked_number = -1; int picked_number = -1;
for (size_t i = 0; i < filenames.size(); i++) { for (size_t i = 0; i < filenames.size(); i++) {
if (ParseFileName(filenames[i], &number, &type) && type == filetype && if (ParseFileName(filenames[i], &number, &type) &&
type == filetype &&
int(number) > picked_number) { // Pick latest file int(number) > picked_number) { // Pick latest file
fname = dbname_ + "/" + filenames[i]; fname = dbname_ + "/" + filenames[i];
picked_number = number; picked_number = number;
@ -127,32 +141,35 @@ class CorruptionTest : public testing::Test {
} }
ASSERT_TRUE(!fname.empty()) << filetype; ASSERT_TRUE(!fname.empty()) << filetype;
uint64_t file_size; struct stat sbuf;
ASSERT_LEVELDB_OK(env_.target()->GetFileSize(fname, &file_size)); if (stat(fname.c_str(), &sbuf) != 0) {
const char* msg = strerror(errno);
ASSERT_TRUE(false) << fname << ": " << msg;
}
if (offset < 0) { if (offset < 0) {
// Relative to end of file; make it absolute // Relative to end of file; make it absolute
if (-offset > file_size) { if (-offset > sbuf.st_size) {
offset = 0; offset = 0;
} else { } else {
offset = file_size + offset; offset = sbuf.st_size + offset;
} }
} }
if (offset > file_size) { if (offset > sbuf.st_size) {
offset = file_size; offset = sbuf.st_size;
} }
if (offset + bytes_to_corrupt > file_size) { if (offset + bytes_to_corrupt > sbuf.st_size) {
bytes_to_corrupt = file_size - offset; bytes_to_corrupt = sbuf.st_size - offset;
} }
// Do it // Do it
std::string contents; std::string contents;
Status s = ReadFileToString(env_.target(), fname, &contents); Status s = ReadFileToString(Env::Default(), fname, &contents);
ASSERT_TRUE(s.ok()) << s.ToString(); ASSERT_TRUE(s.ok()) << s.ToString();
for (int i = 0; i < bytes_to_corrupt; i++) { for (int i = 0; i < bytes_to_corrupt; i++) {
contents[i + offset] ^= 0x80; contents[i + offset] ^= 0x80;
} }
s = WriteStringToFile(env_.target(), contents, fname); s = WriteStringToFile(Env::Default(), contents, fname);
ASSERT_TRUE(s.ok()) << s.ToString(); ASSERT_TRUE(s.ok()) << s.ToString();
} }
@ -170,7 +187,7 @@ class CorruptionTest : public testing::Test {
// Return the ith key // Return the ith key
Slice Key(int i, std::string* storage) { Slice Key(int i, std::string* storage) {
char buf[100]; char buf[100];
std::snprintf(buf, sizeof(buf), "%016d", i); snprintf(buf, sizeof(buf), "%016d", i);
storage->assign(buf, strlen(buf)); storage->assign(buf, strlen(buf));
return Slice(*storage); return Slice(*storage);
} }
@ -180,20 +197,12 @@ class CorruptionTest : public testing::Test {
Random r(k); Random r(k);
return test::RandomString(&r, kValueSize, storage); return test::RandomString(&r, kValueSize, storage);
} }
test::ErrorEnv env_;
Options options_;
DB* db_;
private:
std::string dbname_;
Cache* tiny_cache_;
}; };
TEST_F(CorruptionTest, Recovery) { TEST(CorruptionTest, Recovery) {
Build(100); Build(100);
Check(100, 100); Check(100, 100);
Corrupt(kLogFile, 19, 1); // WriteBatch tag for first record Corrupt(kLogFile, 19, 1); // WriteBatch tag for first record
Corrupt(kLogFile, log::kBlockSize + 1000, 1); // Somewhere in second block Corrupt(kLogFile, log::kBlockSize + 1000, 1); // Somewhere in second block
Reopen(); Reopen();
@ -201,13 +210,13 @@ TEST_F(CorruptionTest, Recovery) {
Check(36, 36); Check(36, 36);
} }
TEST_F(CorruptionTest, RecoverWriteError) { TEST(CorruptionTest, RecoverWriteError) {
env_.writable_file_error_ = true; env_.writable_file_error_ = true;
Status s = TryReopen(); Status s = TryReopen();
ASSERT_TRUE(!s.ok()); ASSERT_TRUE(!s.ok());
} }
TEST_F(CorruptionTest, NewFileErrorDuringWrite) { TEST(CorruptionTest, NewFileErrorDuringWrite) {
// Do enough writing to force minor compaction // Do enough writing to force minor compaction
env_.writable_file_error_ = true; env_.writable_file_error_ = true;
const int num = 3 + (Options().write_buffer_size / kValueSize); const int num = 3 + (Options().write_buffer_size / kValueSize);
@ -224,26 +233,26 @@ TEST_F(CorruptionTest, NewFileErrorDuringWrite) {
Reopen(); Reopen();
} }
TEST_F(CorruptionTest, TableFile) { TEST(CorruptionTest, TableFile) {
Build(100); Build(100);
DBImpl* dbi = reinterpret_cast<DBImpl*>(db_); DBImpl* dbi = reinterpret_cast<DBImpl*>(db_);
dbi->TEST_CompactMemTable(); dbi->TEST_CompactMemTable();
dbi->TEST_CompactRange(0, nullptr, nullptr); dbi->TEST_CompactRange(0, NULL, NULL);
dbi->TEST_CompactRange(1, nullptr, nullptr); dbi->TEST_CompactRange(1, NULL, NULL);
Corrupt(kTableFile, 100, 1); Corrupt(kTableFile, 100, 1);
Check(90, 99); Check(90, 99);
} }
TEST_F(CorruptionTest, TableFileRepair) { TEST(CorruptionTest, TableFileRepair) {
options_.block_size = 2 * kValueSize; // Limit scope of corruption options_.block_size = 2 * kValueSize; // Limit scope of corruption
options_.paranoid_checks = true; options_.paranoid_checks = true;
Reopen(); Reopen();
Build(100); Build(100);
DBImpl* dbi = reinterpret_cast<DBImpl*>(db_); DBImpl* dbi = reinterpret_cast<DBImpl*>(db_);
dbi->TEST_CompactMemTable(); dbi->TEST_CompactMemTable();
dbi->TEST_CompactRange(0, nullptr, nullptr); dbi->TEST_CompactRange(0, NULL, NULL);
dbi->TEST_CompactRange(1, nullptr, nullptr); dbi->TEST_CompactRange(1, NULL, NULL);
Corrupt(kTableFile, 100, 1); Corrupt(kTableFile, 100, 1);
RepairDB(); RepairDB();
@ -251,7 +260,7 @@ TEST_F(CorruptionTest, TableFileRepair) {
Check(95, 99); Check(95, 99);
} }
TEST_F(CorruptionTest, TableFileIndexData) { TEST(CorruptionTest, TableFileIndexData) {
Build(10000); // Enough to build multiple Tables Build(10000); // Enough to build multiple Tables
DBImpl* dbi = reinterpret_cast<DBImpl*>(db_); DBImpl* dbi = reinterpret_cast<DBImpl*>(db_);
dbi->TEST_CompactMemTable(); dbi->TEST_CompactMemTable();
@ -261,39 +270,39 @@ TEST_F(CorruptionTest, TableFileIndexData) {
Check(5000, 9999); Check(5000, 9999);
} }
TEST_F(CorruptionTest, MissingDescriptor) { TEST(CorruptionTest, MissingDescriptor) {
Build(1000); Build(1000);
RepairDB(); RepairDB();
Reopen(); Reopen();
Check(1000, 1000); Check(1000, 1000);
} }
TEST_F(CorruptionTest, SequenceNumberRecovery) { TEST(CorruptionTest, SequenceNumberRecovery) {
ASSERT_LEVELDB_OK(db_->Put(WriteOptions(), "foo", "v1")); ASSERT_OK(db_->Put(WriteOptions(), "foo", "v1"));
ASSERT_LEVELDB_OK(db_->Put(WriteOptions(), "foo", "v2")); ASSERT_OK(db_->Put(WriteOptions(), "foo", "v2"));
ASSERT_LEVELDB_OK(db_->Put(WriteOptions(), "foo", "v3")); ASSERT_OK(db_->Put(WriteOptions(), "foo", "v3"));
ASSERT_LEVELDB_OK(db_->Put(WriteOptions(), "foo", "v4")); ASSERT_OK(db_->Put(WriteOptions(), "foo", "v4"));
ASSERT_LEVELDB_OK(db_->Put(WriteOptions(), "foo", "v5")); ASSERT_OK(db_->Put(WriteOptions(), "foo", "v5"));
RepairDB(); RepairDB();
Reopen(); Reopen();
std::string v; std::string v;
ASSERT_LEVELDB_OK(db_->Get(ReadOptions(), "foo", &v)); ASSERT_OK(db_->Get(ReadOptions(), "foo", &v));
ASSERT_EQ("v5", v); ASSERT_EQ("v5", v);
// Write something. If sequence number was not recovered properly, // Write something. If sequence number was not recovered properly,
// it will be hidden by an earlier write. // it will be hidden by an earlier write.
ASSERT_LEVELDB_OK(db_->Put(WriteOptions(), "foo", "v6")); ASSERT_OK(db_->Put(WriteOptions(), "foo", "v6"));
ASSERT_LEVELDB_OK(db_->Get(ReadOptions(), "foo", &v)); ASSERT_OK(db_->Get(ReadOptions(), "foo", &v));
ASSERT_EQ("v6", v); ASSERT_EQ("v6", v);
Reopen(); Reopen();
ASSERT_LEVELDB_OK(db_->Get(ReadOptions(), "foo", &v)); ASSERT_OK(db_->Get(ReadOptions(), "foo", &v));
ASSERT_EQ("v6", v); ASSERT_EQ("v6", v);
} }
TEST_F(CorruptionTest, CorruptedDescriptor) { TEST(CorruptionTest, CorruptedDescriptor) {
ASSERT_LEVELDB_OK(db_->Put(WriteOptions(), "foo", "hello")); ASSERT_OK(db_->Put(WriteOptions(), "foo", "hello"));
DBImpl* dbi = reinterpret_cast<DBImpl*>(db_); DBImpl* dbi = reinterpret_cast<DBImpl*>(db_);
dbi->TEST_CompactMemTable(); dbi->TEST_CompactMemTable();
dbi->TEST_CompactRange(0, nullptr, nullptr); dbi->TEST_CompactRange(0, NULL, NULL);
Corrupt(kDescriptorFile, 0, 1000); Corrupt(kDescriptorFile, 0, 1000);
Status s = TryReopen(); Status s = TryReopen();
@ -302,11 +311,11 @@ TEST_F(CorruptionTest, CorruptedDescriptor) {
RepairDB(); RepairDB();
Reopen(); Reopen();
std::string v; std::string v;
ASSERT_LEVELDB_OK(db_->Get(ReadOptions(), "foo", &v)); ASSERT_OK(db_->Get(ReadOptions(), "foo", &v));
ASSERT_EQ("hello", v); ASSERT_EQ("hello", v);
} }
TEST_F(CorruptionTest, CompactionInputError) { TEST(CorruptionTest, CompactionInputError) {
Build(10); Build(10);
DBImpl* dbi = reinterpret_cast<DBImpl*>(db_); DBImpl* dbi = reinterpret_cast<DBImpl*>(db_);
dbi->TEST_CompactMemTable(); dbi->TEST_CompactMemTable();
@ -321,7 +330,7 @@ TEST_F(CorruptionTest, CompactionInputError) {
Check(10000, 10000); Check(10000, 10000);
} }
TEST_F(CorruptionTest, CompactionInputErrorParanoid) { TEST(CorruptionTest, CompactionInputErrorParanoid) {
options_.paranoid_checks = true; options_.paranoid_checks = true;
options_.write_buffer_size = 512 << 10; options_.write_buffer_size = 512 << 10;
Reopen(); Reopen();
@ -334,7 +343,7 @@ TEST_F(CorruptionTest, CompactionInputErrorParanoid) {
Corrupt(kTableFile, 100, 1); Corrupt(kTableFile, 100, 1);
env_.SleepForMicroseconds(100000); env_.SleepForMicroseconds(100000);
} }
dbi->CompactRange(nullptr, nullptr); dbi->CompactRange(NULL, NULL);
// Write must fail because of corrupted table // Write must fail because of corrupted table
std::string tmp1, tmp2; std::string tmp1, tmp2;
@ -342,21 +351,24 @@ TEST_F(CorruptionTest, CompactionInputErrorParanoid) {
ASSERT_TRUE(!s.ok()) << "write did not fail in corrupted paranoid db"; ASSERT_TRUE(!s.ok()) << "write did not fail in corrupted paranoid db";
} }
TEST_F(CorruptionTest, UnrelatedKeys) { TEST(CorruptionTest, UnrelatedKeys) {
Build(10); Build(10);
DBImpl* dbi = reinterpret_cast<DBImpl*>(db_); DBImpl* dbi = reinterpret_cast<DBImpl*>(db_);
dbi->TEST_CompactMemTable(); dbi->TEST_CompactMemTable();
Corrupt(kTableFile, 100, 1); Corrupt(kTableFile, 100, 1);
std::string tmp1, tmp2; std::string tmp1, tmp2;
ASSERT_LEVELDB_OK( ASSERT_OK(db_->Put(WriteOptions(), Key(1000, &tmp1), Value(1000, &tmp2)));
db_->Put(WriteOptions(), Key(1000, &tmp1), Value(1000, &tmp2)));
std::string v; std::string v;
ASSERT_LEVELDB_OK(db_->Get(ReadOptions(), Key(1000, &tmp1), &v)); ASSERT_OK(db_->Get(ReadOptions(), Key(1000, &tmp1), &v));
ASSERT_EQ(Value(1000, &tmp2).ToString(), v); ASSERT_EQ(Value(1000, &tmp2).ToString(), v);
dbi->TEST_CompactMemTable(); dbi->TEST_CompactMemTable();
ASSERT_LEVELDB_OK(db_->Get(ReadOptions(), Key(1000, &tmp1), &v)); ASSERT_OK(db_->Get(ReadOptions(), Key(1000, &tmp1), &v));
ASSERT_EQ(Value(1000, &tmp2).ToString(), v); ASSERT_EQ(Value(1000, &tmp2).ToString(), v);
} }
} // namespace leveldb } // namespace leveldb
int main(int argc, char** argv) {
return leveldb::test::RunAllTests();
}

File diff suppressed because it is too large Load Diff

File diff suppressed because it is too large Load Diff

View File

@ -5,11 +5,8 @@
#ifndef STORAGE_LEVELDB_DB_DB_IMPL_H_ #ifndef STORAGE_LEVELDB_DB_DB_IMPL_H_
#define STORAGE_LEVELDB_DB_DB_IMPL_H_ #define STORAGE_LEVELDB_DB_DB_IMPL_H_
#include <atomic>
#include <deque> #include <deque>
#include <set> #include <set>
#include <string>
#include "db/dbformat.h" #include "db/dbformat.h"
#include "db/log_writer.h" #include "db/log_writer.h"
#include "db/snapshot.h" #include "db/snapshot.h"
@ -29,25 +26,21 @@ class VersionSet;
class DBImpl : public DB { class DBImpl : public DB {
public: public:
DBImpl(const Options& options, const std::string& dbname); DBImpl(const Options& options, const std::string& dbname);
virtual ~DBImpl();
DBImpl(const DBImpl&) = delete;
DBImpl& operator=(const DBImpl&) = delete;
~DBImpl() override;
// Implementations of the DB interface // Implementations of the DB interface
Status Put(const WriteOptions&, const Slice& key, virtual Status Put(const WriteOptions&, const Slice& key, const Slice& value);
const Slice& value) override; virtual Status Delete(const WriteOptions&, const Slice& key);
Status Delete(const WriteOptions&, const Slice& key) override; virtual Status Write(const WriteOptions& options, WriteBatch* updates);
Status Write(const WriteOptions& options, WriteBatch* updates) override; virtual Status Get(const ReadOptions& options,
Status Get(const ReadOptions& options, const Slice& key, const Slice& key,
std::string* value) override; std::string* value);
Iterator* NewIterator(const ReadOptions&) override; virtual Iterator* NewIterator(const ReadOptions&);
const Snapshot* GetSnapshot() override; virtual const Snapshot* GetSnapshot();
void ReleaseSnapshot(const Snapshot* snapshot) override; virtual void ReleaseSnapshot(const Snapshot* snapshot);
bool GetProperty(const Slice& property, std::string* value) override; virtual bool GetProperty(const Slice& property, std::string* value);
void GetApproximateSizes(const Range* range, int n, uint64_t* sizes) override; virtual void GetApproximateSizes(const Range* range, int n, uint64_t* sizes);
void CompactRange(const Slice* begin, const Slice* end) override; virtual void CompactRange(const Slice* begin, const Slice* end);
// Extra methods (for testing) that are not in the public DB interface // Extra methods (for testing) that are not in the public DB interface
@ -76,31 +69,6 @@ class DBImpl : public DB {
struct CompactionState; struct CompactionState;
struct Writer; struct Writer;
// Information for a manual compaction
struct ManualCompaction {
int level;
bool done;
const InternalKey* begin; // null means beginning of key range
const InternalKey* end; // null means end of key range
InternalKey tmp_storage; // Used to keep track of compaction progress
};
// Per level compaction stats. stats_[level] stores the stats for
// compactions that produced data for the specified "level".
struct CompactionStats {
CompactionStats() : micros(0), bytes_read(0), bytes_written(0) {}
void Add(const CompactionStats& c) {
this->micros += c.micros;
this->bytes_read += c.bytes_read;
this->bytes_written += c.bytes_written;
}
int64_t micros;
int64_t bytes_read;
int64_t bytes_written;
};
Iterator* NewInternalIterator(const ReadOptions&, Iterator* NewInternalIterator(const ReadOptions&,
SequenceNumber* latest_snapshot, SequenceNumber* latest_snapshot,
uint32_t* seed); uint32_t* seed);
@ -116,7 +84,7 @@ class DBImpl : public DB {
void MaybeIgnoreError(Status* s) const; void MaybeIgnoreError(Status* s) const;
// Delete any unneeded files and stale in-memory entries. // Delete any unneeded files and stale in-memory entries.
void RemoveObsoleteFiles() EXCLUSIVE_LOCKS_REQUIRED(mutex_); void DeleteObsoleteFiles();
// Compact the in-memory write buffer to disk. Switches to a new // Compact the in-memory write buffer to disk. Switches to a new
// log-file/memtable and writes a new descriptor iff successful. // log-file/memtable and writes a new descriptor iff successful.
@ -132,15 +100,14 @@ class DBImpl : public DB {
Status MakeRoomForWrite(bool force /* compact even if there is room? */) Status MakeRoomForWrite(bool force /* compact even if there is room? */)
EXCLUSIVE_LOCKS_REQUIRED(mutex_); EXCLUSIVE_LOCKS_REQUIRED(mutex_);
WriteBatch* BuildBatchGroup(Writer** last_writer) WriteBatch* BuildBatchGroup(Writer** last_writer);
EXCLUSIVE_LOCKS_REQUIRED(mutex_);
void RecordBackgroundError(const Status& s); void RecordBackgroundError(const Status& s);
void MaybeScheduleCompaction() EXCLUSIVE_LOCKS_REQUIRED(mutex_); void MaybeScheduleCompaction() EXCLUSIVE_LOCKS_REQUIRED(mutex_);
static void BGWork(void* db); static void BGWork(void* db);
void BackgroundCall(); void BackgroundCall();
void BackgroundCompaction() EXCLUSIVE_LOCKS_REQUIRED(mutex_); void BackgroundCompaction() EXCLUSIVE_LOCKS_REQUIRED(mutex_);
void CleanupCompaction(CompactionState* compact) void CleanupCompaction(CompactionState* compact)
EXCLUSIVE_LOCKS_REQUIRED(mutex_); EXCLUSIVE_LOCKS_REQUIRED(mutex_);
Status DoCompactionWork(CompactionState* compact) Status DoCompactionWork(CompactionState* compact)
@ -151,66 +118,93 @@ class DBImpl : public DB {
Status InstallCompactionResults(CompactionState* compact) Status InstallCompactionResults(CompactionState* compact)
EXCLUSIVE_LOCKS_REQUIRED(mutex_); EXCLUSIVE_LOCKS_REQUIRED(mutex_);
const Comparator* user_comparator() const {
return internal_comparator_.user_comparator();
}
// Constant after construction // Constant after construction
Env* const env_; Env* const env_;
const InternalKeyComparator internal_comparator_; const InternalKeyComparator internal_comparator_;
const InternalFilterPolicy internal_filter_policy_; const InternalFilterPolicy internal_filter_policy_;
const Options options_; // options_.comparator == &internal_comparator_ const Options options_; // options_.comparator == &internal_comparator_
const bool owns_info_log_; bool owns_info_log_;
const bool owns_cache_; bool owns_cache_;
const std::string dbname_; const std::string dbname_;
// table_cache_ provides its own synchronization // table_cache_ provides its own synchronization
TableCache* const table_cache_; TableCache* table_cache_;
// Lock over the persistent DB state. Non-null iff successfully acquired. // Lock over the persistent DB state. Non-NULL iff successfully acquired.
FileLock* db_lock_; FileLock* db_lock_;
// State below is protected by mutex_ // State below is protected by mutex_
port::Mutex mutex_; port::Mutex mutex_;
std::atomic<bool> shutting_down_; port::AtomicPointer shutting_down_;
port::CondVar background_work_finished_signal_ GUARDED_BY(mutex_); port::CondVar bg_cv_; // Signalled when background work finishes
MemTable* mem_; MemTable* mem_;
MemTable* imm_ GUARDED_BY(mutex_); // Memtable being compacted MemTable* imm_; // Memtable being compacted
std::atomic<bool> has_imm_; // So bg thread can detect non-null imm_ port::AtomicPointer has_imm_; // So bg thread can detect non-NULL imm_
WritableFile* logfile_; WritableFile* logfile_;
uint64_t logfile_number_ GUARDED_BY(mutex_); uint64_t logfile_number_;
log::Writer* log_; log::Writer* log_;
uint32_t seed_ GUARDED_BY(mutex_); // For sampling. uint32_t seed_; // For sampling.
// Queue of writers. // Queue of writers.
std::deque<Writer*> writers_ GUARDED_BY(mutex_); std::deque<Writer*> writers_;
WriteBatch* tmp_batch_ GUARDED_BY(mutex_); WriteBatch* tmp_batch_;
SnapshotList snapshots_ GUARDED_BY(mutex_); SnapshotList snapshots_;
// Set of table files to protect from deletion because they are // Set of table files to protect from deletion because they are
// part of ongoing compactions. // part of ongoing compactions.
std::set<uint64_t> pending_outputs_ GUARDED_BY(mutex_); std::set<uint64_t> pending_outputs_;
// Has a background compaction been scheduled or is running? // Has a background compaction been scheduled or is running?
bool background_compaction_scheduled_ GUARDED_BY(mutex_); bool bg_compaction_scheduled_;
ManualCompaction* manual_compaction_ GUARDED_BY(mutex_); // Information for a manual compaction
struct ManualCompaction {
int level;
bool done;
const InternalKey* begin; // NULL means beginning of key range
const InternalKey* end; // NULL means end of key range
InternalKey tmp_storage; // Used to keep track of compaction progress
};
ManualCompaction* manual_compaction_;
VersionSet* const versions_ GUARDED_BY(mutex_); VersionSet* versions_;
// Have we encountered a background error in paranoid mode? // Have we encountered a background error in paranoid mode?
Status bg_error_ GUARDED_BY(mutex_); Status bg_error_;
CompactionStats stats_[config::kNumLevels] GUARDED_BY(mutex_); // Per level compaction stats. stats_[level] stores the stats for
// compactions that produced data for the specified "level".
struct CompactionStats {
int64_t micros;
int64_t bytes_read;
int64_t bytes_written;
CompactionStats() : micros(0), bytes_read(0), bytes_written(0) { }
void Add(const CompactionStats& c) {
this->micros += c.micros;
this->bytes_read += c.bytes_read;
this->bytes_written += c.bytes_written;
}
};
CompactionStats stats_[config::kNumLevels];
// No copying allowed
DBImpl(const DBImpl&);
void operator=(const DBImpl&);
const Comparator* user_comparator() const {
return internal_comparator_.user_comparator();
}
}; };
// Sanitize db options. The caller should delete result.info_log if // Sanitize db options. The caller should delete result.info_log if
// it is not equal to src.info_log. // it is not equal to src.info_log.
Options SanitizeOptions(const std::string& db, extern Options SanitizeOptions(const std::string& db,
const InternalKeyComparator* icmp, const InternalKeyComparator* icmp,
const InternalFilterPolicy* ipolicy, const InternalFilterPolicy* ipolicy,
const Options& src); const Options& src);
} // namespace leveldb } // namespace leveldb

View File

@ -4,9 +4,9 @@
#include "db/db_iter.h" #include "db/db_iter.h"
#include "db/filename.h"
#include "db/db_impl.h" #include "db/db_impl.h"
#include "db/dbformat.h" #include "db/dbformat.h"
#include "db/filename.h"
#include "leveldb/env.h" #include "leveldb/env.h"
#include "leveldb/iterator.h" #include "leveldb/iterator.h"
#include "port/port.h" #include "port/port.h"
@ -21,9 +21,9 @@ static void DumpInternalIter(Iterator* iter) {
for (iter->SeekToFirst(); iter->Valid(); iter->Next()) { for (iter->SeekToFirst(); iter->Valid(); iter->Next()) {
ParsedInternalKey k; ParsedInternalKey k;
if (!ParseInternalKey(iter->key(), &k)) { if (!ParseInternalKey(iter->key(), &k)) {
std::fprintf(stderr, "Corrupt '%s'\n", EscapeString(iter->key()).c_str()); fprintf(stderr, "Corrupt '%s'\n", EscapeString(iter->key()).c_str());
} else { } else {
std::fprintf(stderr, "@ '%s'\n", k.DebugString().c_str()); fprintf(stderr, "@ '%s'\n", k.DebugString().c_str());
} }
} }
} }
@ -36,14 +36,17 @@ namespace {
// combines multiple entries for the same userkey found in the DB // combines multiple entries for the same userkey found in the DB
// representation into a single entry while accounting for sequence // representation into a single entry while accounting for sequence
// numbers, deletion markers, overwrites, etc. // numbers, deletion markers, overwrites, etc.
class DBIter : public Iterator { class DBIter: public Iterator {
public: public:
// Which direction is the iterator currently moving? // Which direction is the iterator currently moving?
// (1) When moving forward, the internal iterator is positioned at // (1) When moving forward, the internal iterator is positioned at
// the exact entry that yields this->key(), this->value() // the exact entry that yields this->key(), this->value()
// (2) When moving backwards, the internal iterator is positioned // (2) When moving backwards, the internal iterator is positioned
// just before all entries whose user key == this->key(). // just before all entries whose user key == this->key().
enum Direction { kForward, kReverse }; enum Direction {
kForward,
kReverse
};
DBIter(DBImpl* db, const Comparator* cmp, Iterator* iter, SequenceNumber s, DBIter(DBImpl* db, const Comparator* cmp, Iterator* iter, SequenceNumber s,
uint32_t seed) uint32_t seed)
@ -54,22 +57,21 @@ class DBIter : public Iterator {
direction_(kForward), direction_(kForward),
valid_(false), valid_(false),
rnd_(seed), rnd_(seed),
bytes_until_read_sampling_(RandomCompactionPeriod()) {} bytes_counter_(RandomPeriod()) {
}
DBIter(const DBIter&) = delete; virtual ~DBIter() {
DBIter& operator=(const DBIter&) = delete; delete iter_;
}
~DBIter() override { delete iter_; } virtual bool Valid() const { return valid_; }
bool Valid() const override { return valid_; } virtual Slice key() const {
Slice key() const override {
assert(valid_); assert(valid_);
return (direction_ == kForward) ? ExtractUserKey(iter_->key()) : saved_key_; return (direction_ == kForward) ? ExtractUserKey(iter_->key()) : saved_key_;
} }
Slice value() const override { virtual Slice value() const {
assert(valid_); assert(valid_);
return (direction_ == kForward) ? iter_->value() : saved_value_; return (direction_ == kForward) ? iter_->value() : saved_value_;
} }
Status status() const override { virtual Status status() const {
if (status_.ok()) { if (status_.ok()) {
return iter_->status(); return iter_->status();
} else { } else {
@ -77,11 +79,11 @@ class DBIter : public Iterator {
} }
} }
void Next() override; virtual void Next();
void Prev() override; virtual void Prev();
void Seek(const Slice& target) override; virtual void Seek(const Slice& target);
void SeekToFirst() override; virtual void SeekToFirst();
void SeekToLast() override; virtual void SeekToLast();
private: private:
void FindNextUserEntry(bool skipping, std::string* skip); void FindNextUserEntry(bool skipping, std::string* skip);
@ -101,35 +103,38 @@ class DBIter : public Iterator {
} }
} }
// Picks the number of bytes that can be read until a compaction is scheduled. // Pick next gap with average value of config::kReadBytesPeriod.
size_t RandomCompactionPeriod() { ssize_t RandomPeriod() {
return rnd_.Uniform(2 * config::kReadBytesPeriod); return rnd_.Uniform(2*config::kReadBytesPeriod);
} }
DBImpl* db_; DBImpl* db_;
const Comparator* const user_comparator_; const Comparator* const user_comparator_;
Iterator* const iter_; Iterator* const iter_;
SequenceNumber const sequence_; SequenceNumber const sequence_;
Status status_; Status status_;
std::string saved_key_; // == current key when direction_==kReverse std::string saved_key_; // == current key when direction_==kReverse
std::string saved_value_; // == current raw value when direction_==kReverse std::string saved_value_; // == current raw value when direction_==kReverse
Direction direction_; Direction direction_;
bool valid_; bool valid_;
Random rnd_; Random rnd_;
size_t bytes_until_read_sampling_; ssize_t bytes_counter_;
// No copying allowed
DBIter(const DBIter&);
void operator=(const DBIter&);
}; };
inline bool DBIter::ParseKey(ParsedInternalKey* ikey) { inline bool DBIter::ParseKey(ParsedInternalKey* ikey) {
Slice k = iter_->key(); Slice k = iter_->key();
ssize_t n = k.size() + iter_->value().size();
size_t bytes_read = k.size() + iter_->value().size(); bytes_counter_ -= n;
while (bytes_until_read_sampling_ < bytes_read) { while (bytes_counter_ < 0) {
bytes_until_read_sampling_ += RandomCompactionPeriod(); bytes_counter_ += RandomPeriod();
db_->RecordReadSample(k); db_->RecordReadSample(k);
} }
assert(bytes_until_read_sampling_ >= bytes_read);
bytes_until_read_sampling_ -= bytes_read;
if (!ParseInternalKey(k, ikey)) { if (!ParseInternalKey(k, ikey)) {
status_ = Status::Corruption("corrupted internal key in DBIter"); status_ = Status::Corruption("corrupted internal key in DBIter");
return false; return false;
@ -160,15 +165,6 @@ void DBIter::Next() {
} else { } else {
// Store in saved_key_ the current key so we skip it below. // Store in saved_key_ the current key so we skip it below.
SaveKey(ExtractUserKey(iter_->key()), &saved_key_); SaveKey(ExtractUserKey(iter_->key()), &saved_key_);
// iter_ is pointing to current key. We can now safely move to the next to
// avoid checking current key.
iter_->Next();
if (!iter_->Valid()) {
valid_ = false;
saved_key_.clear();
return;
}
} }
FindNextUserEntry(true, &saved_key_); FindNextUserEntry(true, &saved_key_);
@ -222,8 +218,8 @@ void DBIter::Prev() {
ClearSavedValue(); ClearSavedValue();
return; return;
} }
if (user_comparator_->Compare(ExtractUserKey(iter_->key()), saved_key_) < if (user_comparator_->Compare(ExtractUserKey(iter_->key()),
0) { saved_key_) < 0) {
break; break;
} }
} }
@ -279,8 +275,8 @@ void DBIter::Seek(const Slice& target) {
direction_ = kForward; direction_ = kForward;
ClearSavedValue(); ClearSavedValue();
saved_key_.clear(); saved_key_.clear();
AppendInternalKey(&saved_key_, AppendInternalKey(
ParsedInternalKey(target, sequence_, kValueTypeForSeek)); &saved_key_, ParsedInternalKey(target, sequence_, kValueTypeForSeek));
iter_->Seek(saved_key_); iter_->Seek(saved_key_);
if (iter_->Valid()) { if (iter_->Valid()) {
FindNextUserEntry(false, &saved_key_ /* temporary storage */); FindNextUserEntry(false, &saved_key_ /* temporary storage */);
@ -309,9 +305,12 @@ void DBIter::SeekToLast() {
} // anonymous namespace } // anonymous namespace
Iterator* NewDBIterator(DBImpl* db, const Comparator* user_key_comparator, Iterator* NewDBIterator(
Iterator* internal_iter, SequenceNumber sequence, DBImpl* db,
uint32_t seed) { const Comparator* user_key_comparator,
Iterator* internal_iter,
SequenceNumber sequence,
uint32_t seed) {
return new DBIter(db, user_key_comparator, internal_iter, sequence, seed); return new DBIter(db, user_key_comparator, internal_iter, sequence, seed);
} }

View File

@ -5,10 +5,9 @@
#ifndef STORAGE_LEVELDB_DB_DB_ITER_H_ #ifndef STORAGE_LEVELDB_DB_DB_ITER_H_
#define STORAGE_LEVELDB_DB_DB_ITER_H_ #define STORAGE_LEVELDB_DB_DB_ITER_H_
#include <cstdint> #include <stdint.h>
#include "db/dbformat.h"
#include "leveldb/db.h" #include "leveldb/db.h"
#include "db/dbformat.h"
namespace leveldb { namespace leveldb {
@ -17,9 +16,12 @@ class DBImpl;
// Return a new iterator that converts internal keys (yielded by // Return a new iterator that converts internal keys (yielded by
// "*internal_iter") that were live at the specified "sequence" number // "*internal_iter") that were live at the specified "sequence" number
// into appropriate user keys. // into appropriate user keys.
Iterator* NewDBIterator(DBImpl* db, const Comparator* user_key_comparator, extern Iterator* NewDBIterator(
Iterator* internal_iter, SequenceNumber sequence, DBImpl* db,
uint32_t seed); const Comparator* user_key_comparator,
Iterator* internal_iter,
SequenceNumber sequence,
uint32_t seed);
} // namespace leveldb } // namespace leveldb

File diff suppressed because it is too large Load Diff

View File

@ -2,11 +2,8 @@
// Use of this source code is governed by a BSD-style license that can be // Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file. See the AUTHORS file for names of contributors. // found in the LICENSE file. See the AUTHORS file for names of contributors.
#include <stdio.h>
#include "db/dbformat.h" #include "db/dbformat.h"
#include <cstdio>
#include <sstream>
#include "port/port.h" #include "port/port.h"
#include "util/coding.h" #include "util/coding.h"
@ -24,20 +21,26 @@ void AppendInternalKey(std::string* result, const ParsedInternalKey& key) {
} }
std::string ParsedInternalKey::DebugString() const { std::string ParsedInternalKey::DebugString() const {
std::ostringstream ss; char buf[50];
ss << '\'' << EscapeString(user_key.ToString()) << "' @ " << sequence << " : " snprintf(buf, sizeof(buf), "' @ %llu : %d",
<< static_cast<int>(type); (unsigned long long) sequence,
return ss.str(); int(type));
std::string result = "'";
result += EscapeString(user_key.ToString());
result += buf;
return result;
} }
std::string InternalKey::DebugString() const { std::string InternalKey::DebugString() const {
std::string result;
ParsedInternalKey parsed; ParsedInternalKey parsed;
if (ParseInternalKey(rep_, &parsed)) { if (ParseInternalKey(rep_, &parsed)) {
return parsed.DebugString(); result = parsed.DebugString();
} else {
result = "(bad)";
result.append(EscapeString(rep_));
} }
std::ostringstream ss; return result;
ss << "(bad)" << EscapeString(rep_);
return ss.str();
} }
const char* InternalKeyComparator::Name() const { const char* InternalKeyComparator::Name() const {
@ -62,8 +65,9 @@ int InternalKeyComparator::Compare(const Slice& akey, const Slice& bkey) const {
return r; return r;
} }
void InternalKeyComparator::FindShortestSeparator(std::string* start, void InternalKeyComparator::FindShortestSeparator(
const Slice& limit) const { std::string* start,
const Slice& limit) const {
// Attempt to shorten the user portion of the key // Attempt to shorten the user portion of the key
Slice user_start = ExtractUserKey(*start); Slice user_start = ExtractUserKey(*start);
Slice user_limit = ExtractUserKey(limit); Slice user_limit = ExtractUserKey(limit);
@ -73,8 +77,7 @@ void InternalKeyComparator::FindShortestSeparator(std::string* start,
user_comparator_->Compare(user_start, tmp) < 0) { user_comparator_->Compare(user_start, tmp) < 0) {
// User key has become shorter physically, but larger logically. // User key has become shorter physically, but larger logically.
// Tack on the earliest possible number to the shortened user key. // Tack on the earliest possible number to the shortened user key.
PutFixed64(&tmp, PutFixed64(&tmp, PackSequenceAndType(kMaxSequenceNumber,kValueTypeForSeek));
PackSequenceAndType(kMaxSequenceNumber, kValueTypeForSeek));
assert(this->Compare(*start, tmp) < 0); assert(this->Compare(*start, tmp) < 0);
assert(this->Compare(tmp, limit) < 0); assert(this->Compare(tmp, limit) < 0);
start->swap(tmp); start->swap(tmp);
@ -89,14 +92,15 @@ void InternalKeyComparator::FindShortSuccessor(std::string* key) const {
user_comparator_->Compare(user_key, tmp) < 0) { user_comparator_->Compare(user_key, tmp) < 0) {
// User key has become shorter physically, but larger logically. // User key has become shorter physically, but larger logically.
// Tack on the earliest possible number to the shortened user key. // Tack on the earliest possible number to the shortened user key.
PutFixed64(&tmp, PutFixed64(&tmp, PackSequenceAndType(kMaxSequenceNumber,kValueTypeForSeek));
PackSequenceAndType(kMaxSequenceNumber, kValueTypeForSeek));
assert(this->Compare(*key, tmp) < 0); assert(this->Compare(*key, tmp) < 0);
key->swap(tmp); key->swap(tmp);
} }
} }
const char* InternalFilterPolicy::Name() const { return user_policy_->Name(); } const char* InternalFilterPolicy::Name() const {
return user_policy_->Name();
}
void InternalFilterPolicy::CreateFilter(const Slice* keys, int n, void InternalFilterPolicy::CreateFilter(const Slice* keys, int n,
std::string* dst) const { std::string* dst) const {
@ -126,7 +130,7 @@ LookupKey::LookupKey(const Slice& user_key, SequenceNumber s) {
start_ = dst; start_ = dst;
dst = EncodeVarint32(dst, usize + 8); dst = EncodeVarint32(dst, usize + 8);
kstart_ = dst; kstart_ = dst;
std::memcpy(dst, user_key.data(), usize); memcpy(dst, user_key.data(), usize);
dst += usize; dst += usize;
EncodeFixed64(dst, PackSequenceAndType(s, kValueTypeForSeek)); EncodeFixed64(dst, PackSequenceAndType(s, kValueTypeForSeek));
dst += 8; dst += 8;

View File

@ -5,10 +5,7 @@
#ifndef STORAGE_LEVELDB_DB_DBFORMAT_H_ #ifndef STORAGE_LEVELDB_DB_DBFORMAT_H_
#define STORAGE_LEVELDB_DB_DBFORMAT_H_ #define STORAGE_LEVELDB_DB_DBFORMAT_H_
#include <cstddef> #include <stdio.h>
#include <cstdint>
#include <string>
#include "leveldb/comparator.h" #include "leveldb/comparator.h"
#include "leveldb/db.h" #include "leveldb/db.h"
#include "leveldb/filter_policy.h" #include "leveldb/filter_policy.h"
@ -51,7 +48,10 @@ class InternalKey;
// Value types encoded as the last component of internal keys. // Value types encoded as the last component of internal keys.
// DO NOT CHANGE THESE ENUM VALUES: they are embedded in the on-disk // DO NOT CHANGE THESE ENUM VALUES: they are embedded in the on-disk
// data structures. // data structures.
enum ValueType { kTypeDeletion = 0x0, kTypeValue = 0x1 }; enum ValueType {
kTypeDeletion = 0x0,
kTypeValue = 0x1
};
// kValueTypeForSeek defines the ValueType that should be passed when // kValueTypeForSeek defines the ValueType that should be passed when
// constructing a ParsedInternalKey object for seeking to a particular // constructing a ParsedInternalKey object for seeking to a particular
// sequence number (since we sort sequence numbers in decreasing order // sequence number (since we sort sequence numbers in decreasing order
@ -64,16 +64,17 @@ typedef uint64_t SequenceNumber;
// We leave eight bits empty at the bottom so a type and sequence# // We leave eight bits empty at the bottom so a type and sequence#
// can be packed together into 64-bits. // can be packed together into 64-bits.
static const SequenceNumber kMaxSequenceNumber = ((0x1ull << 56) - 1); static const SequenceNumber kMaxSequenceNumber =
((0x1ull << 56) - 1);
struct ParsedInternalKey { struct ParsedInternalKey {
Slice user_key; Slice user_key;
SequenceNumber sequence; SequenceNumber sequence;
ValueType type; ValueType type;
ParsedInternalKey() {} // Intentionally left uninitialized (for speed) ParsedInternalKey() { } // Intentionally left uninitialized (for speed)
ParsedInternalKey(const Slice& u, const SequenceNumber& seq, ValueType t) ParsedInternalKey(const Slice& u, const SequenceNumber& seq, ValueType t)
: user_key(u), sequence(seq), type(t) {} : user_key(u), sequence(seq), type(t) { }
std::string DebugString() const; std::string DebugString() const;
}; };
@ -83,13 +84,15 @@ inline size_t InternalKeyEncodingLength(const ParsedInternalKey& key) {
} }
// Append the serialization of "key" to *result. // Append the serialization of "key" to *result.
void AppendInternalKey(std::string* result, const ParsedInternalKey& key); extern void AppendInternalKey(std::string* result,
const ParsedInternalKey& key);
// Attempt to parse an internal key from "internal_key". On success, // Attempt to parse an internal key from "internal_key". On success,
// stores the parsed data in "*result", and returns true. // stores the parsed data in "*result", and returns true.
// //
// On error, returns false, leaves "*result" in an undefined state. // On error, returns false, leaves "*result" in an undefined state.
bool ParseInternalKey(const Slice& internal_key, ParsedInternalKey* result); extern bool ParseInternalKey(const Slice& internal_key,
ParsedInternalKey* result);
// Returns the user key portion of an internal key. // Returns the user key portion of an internal key.
inline Slice ExtractUserKey(const Slice& internal_key) { inline Slice ExtractUserKey(const Slice& internal_key) {
@ -97,19 +100,27 @@ inline Slice ExtractUserKey(const Slice& internal_key) {
return Slice(internal_key.data(), internal_key.size() - 8); return Slice(internal_key.data(), internal_key.size() - 8);
} }
inline ValueType ExtractValueType(const Slice& internal_key) {
assert(internal_key.size() >= 8);
const size_t n = internal_key.size();
uint64_t num = DecodeFixed64(internal_key.data() + n - 8);
unsigned char c = num & 0xff;
return static_cast<ValueType>(c);
}
// A comparator for internal keys that uses a specified comparator for // A comparator for internal keys that uses a specified comparator for
// the user key portion and breaks ties by decreasing sequence number. // the user key portion and breaks ties by decreasing sequence number.
class InternalKeyComparator : public Comparator { class InternalKeyComparator : public Comparator {
private: private:
const Comparator* user_comparator_; const Comparator* user_comparator_;
public: public:
explicit InternalKeyComparator(const Comparator* c) : user_comparator_(c) {} explicit InternalKeyComparator(const Comparator* c) : user_comparator_(c) { }
const char* Name() const override; virtual const char* Name() const;
int Compare(const Slice& a, const Slice& b) const override; virtual int Compare(const Slice& a, const Slice& b) const;
void FindShortestSeparator(std::string* start, virtual void FindShortestSeparator(
const Slice& limit) const override; std::string* start,
void FindShortSuccessor(std::string* key) const override; const Slice& limit) const;
virtual void FindShortSuccessor(std::string* key) const;
const Comparator* user_comparator() const { return user_comparator_; } const Comparator* user_comparator() const { return user_comparator_; }
@ -120,12 +131,11 @@ class InternalKeyComparator : public Comparator {
class InternalFilterPolicy : public FilterPolicy { class InternalFilterPolicy : public FilterPolicy {
private: private:
const FilterPolicy* const user_policy_; const FilterPolicy* const user_policy_;
public: public:
explicit InternalFilterPolicy(const FilterPolicy* p) : user_policy_(p) {} explicit InternalFilterPolicy(const FilterPolicy* p) : user_policy_(p) { }
const char* Name() const override; virtual const char* Name() const;
void CreateFilter(const Slice* keys, int n, std::string* dst) const override; virtual void CreateFilter(const Slice* keys, int n, std::string* dst) const;
bool KeyMayMatch(const Slice& key, const Slice& filter) const override; virtual bool KeyMayMatch(const Slice& key, const Slice& filter) const;
}; };
// Modules in this directory should keep internal keys wrapped inside // Modules in this directory should keep internal keys wrapped inside
@ -134,18 +144,13 @@ class InternalFilterPolicy : public FilterPolicy {
class InternalKey { class InternalKey {
private: private:
std::string rep_; std::string rep_;
public: public:
InternalKey() {} // Leave rep_ as empty to indicate it is invalid InternalKey() { } // Leave rep_ as empty to indicate it is invalid
InternalKey(const Slice& user_key, SequenceNumber s, ValueType t) { InternalKey(const Slice& user_key, SequenceNumber s, ValueType t) {
AppendInternalKey(&rep_, ParsedInternalKey(user_key, s, t)); AppendInternalKey(&rep_, ParsedInternalKey(user_key, s, t));
} }
bool DecodeFrom(const Slice& s) { void DecodeFrom(const Slice& s) { rep_.assign(s.data(), s.size()); }
rep_.assign(s.data(), s.size());
return !rep_.empty();
}
Slice Encode() const { Slice Encode() const {
assert(!rep_.empty()); assert(!rep_.empty());
return rep_; return rep_;
@ -163,8 +168,8 @@ class InternalKey {
std::string DebugString() const; std::string DebugString() const;
}; };
inline int InternalKeyComparator::Compare(const InternalKey& a, inline int InternalKeyComparator::Compare(
const InternalKey& b) const { const InternalKey& a, const InternalKey& b) const {
return Compare(a.Encode(), b.Encode()); return Compare(a.Encode(), b.Encode());
} }
@ -173,11 +178,11 @@ inline bool ParseInternalKey(const Slice& internal_key,
const size_t n = internal_key.size(); const size_t n = internal_key.size();
if (n < 8) return false; if (n < 8) return false;
uint64_t num = DecodeFixed64(internal_key.data() + n - 8); uint64_t num = DecodeFixed64(internal_key.data() + n - 8);
uint8_t c = num & 0xff; unsigned char c = num & 0xff;
result->sequence = num >> 8; result->sequence = num >> 8;
result->type = static_cast<ValueType>(c); result->type = static_cast<ValueType>(c);
result->user_key = Slice(internal_key.data(), n - 8); result->user_key = Slice(internal_key.data(), n - 8);
return (c <= static_cast<uint8_t>(kTypeValue)); return (c <= static_cast<unsigned char>(kTypeValue));
} }
// A helper class useful for DBImpl::Get() // A helper class useful for DBImpl::Get()
@ -187,9 +192,6 @@ class LookupKey {
// the specified sequence number. // the specified sequence number.
LookupKey(const Slice& user_key, SequenceNumber sequence); LookupKey(const Slice& user_key, SequenceNumber sequence);
LookupKey(const LookupKey&) = delete;
LookupKey& operator=(const LookupKey&) = delete;
~LookupKey(); ~LookupKey();
// Return a key suitable for lookup in a MemTable. // Return a key suitable for lookup in a MemTable.
@ -212,7 +214,11 @@ class LookupKey {
const char* start_; const char* start_;
const char* kstart_; const char* kstart_;
const char* end_; const char* end_;
char space_[200]; // Avoid allocation for short keys char space_[200]; // Avoid allocation for short keys
// No copying allowed
LookupKey(const LookupKey&);
void operator=(const LookupKey&);
}; };
inline LookupKey::~LookupKey() { inline LookupKey::~LookupKey() {

View File

@ -3,13 +3,13 @@
// found in the LICENSE file. See the AUTHORS file for names of contributors. // found in the LICENSE file. See the AUTHORS file for names of contributors.
#include "db/dbformat.h" #include "db/dbformat.h"
#include "gtest/gtest.h"
#include "util/logging.h" #include "util/logging.h"
#include "util/testharness.h"
namespace leveldb { namespace leveldb {
static std::string IKey(const std::string& user_key, uint64_t seq, static std::string IKey(const std::string& user_key,
uint64_t seq,
ValueType vt) { ValueType vt) {
std::string encoded; std::string encoded;
AppendInternalKey(&encoded, ParsedInternalKey(user_key, seq, vt)); AppendInternalKey(&encoded, ParsedInternalKey(user_key, seq, vt));
@ -28,7 +28,9 @@ static std::string ShortSuccessor(const std::string& s) {
return result; return result;
} }
static void TestKey(const std::string& key, uint64_t seq, ValueType vt) { static void TestKey(const std::string& key,
uint64_t seq,
ValueType vt) {
std::string encoded = IKey(key, seq, vt); std::string encoded = IKey(key, seq, vt);
Slice in(encoded); Slice in(encoded);
@ -42,20 +44,16 @@ static void TestKey(const std::string& key, uint64_t seq, ValueType vt) {
ASSERT_TRUE(!ParseInternalKey(Slice("bar"), &decoded)); ASSERT_TRUE(!ParseInternalKey(Slice("bar"), &decoded));
} }
class FormatTest { };
TEST(FormatTest, InternalKey_EncodeDecode) { TEST(FormatTest, InternalKey_EncodeDecode) {
const char* keys[] = {"", "k", "hello", "longggggggggggggggggggggg"}; const char* keys[] = { "", "k", "hello", "longggggggggggggggggggggg" };
const uint64_t seq[] = {1, const uint64_t seq[] = {
2, 1, 2, 3,
3, (1ull << 8) - 1, 1ull << 8, (1ull << 8) + 1,
(1ull << 8) - 1, (1ull << 16) - 1, 1ull << 16, (1ull << 16) + 1,
1ull << 8, (1ull << 32) - 1, 1ull << 32, (1ull << 32) + 1
(1ull << 8) + 1, };
(1ull << 16) - 1,
1ull << 16,
(1ull << 16) + 1,
(1ull << 32) - 1,
1ull << 32,
(1ull << 32) + 1};
for (int k = 0; k < sizeof(keys) / sizeof(keys[0]); k++) { for (int k = 0; k < sizeof(keys) / sizeof(keys[0]); k++) {
for (int s = 0; s < sizeof(seq) / sizeof(seq[0]); s++) { for (int s = 0; s < sizeof(seq) / sizeof(seq[0]); s++) {
TestKey(keys[k], seq[s], kTypeValue); TestKey(keys[k], seq[s], kTypeValue);
@ -64,44 +62,40 @@ TEST(FormatTest, InternalKey_EncodeDecode) {
} }
} }
TEST(FormatTest, InternalKey_DecodeFromEmpty) {
InternalKey internal_key;
ASSERT_TRUE(!internal_key.DecodeFrom(""));
}
TEST(FormatTest, InternalKeyShortSeparator) { TEST(FormatTest, InternalKeyShortSeparator) {
// When user keys are same // When user keys are same
ASSERT_EQ(IKey("foo", 100, kTypeValue), ASSERT_EQ(IKey("foo", 100, kTypeValue),
Shorten(IKey("foo", 100, kTypeValue), IKey("foo", 99, kTypeValue))); Shorten(IKey("foo", 100, kTypeValue),
ASSERT_EQ( IKey("foo", 99, kTypeValue)));
IKey("foo", 100, kTypeValue), ASSERT_EQ(IKey("foo", 100, kTypeValue),
Shorten(IKey("foo", 100, kTypeValue), IKey("foo", 101, kTypeValue))); Shorten(IKey("foo", 100, kTypeValue),
ASSERT_EQ( IKey("foo", 101, kTypeValue)));
IKey("foo", 100, kTypeValue), ASSERT_EQ(IKey("foo", 100, kTypeValue),
Shorten(IKey("foo", 100, kTypeValue), IKey("foo", 100, kTypeValue))); Shorten(IKey("foo", 100, kTypeValue),
ASSERT_EQ( IKey("foo", 100, kTypeValue)));
IKey("foo", 100, kTypeValue), ASSERT_EQ(IKey("foo", 100, kTypeValue),
Shorten(IKey("foo", 100, kTypeValue), IKey("foo", 100, kTypeDeletion))); Shorten(IKey("foo", 100, kTypeValue),
IKey("foo", 100, kTypeDeletion)));
// When user keys are misordered // When user keys are misordered
ASSERT_EQ(IKey("foo", 100, kTypeValue), ASSERT_EQ(IKey("foo", 100, kTypeValue),
Shorten(IKey("foo", 100, kTypeValue), IKey("bar", 99, kTypeValue))); Shorten(IKey("foo", 100, kTypeValue),
IKey("bar", 99, kTypeValue)));
// When user keys are different, but correctly ordered // When user keys are different, but correctly ordered
ASSERT_EQ( ASSERT_EQ(IKey("g", kMaxSequenceNumber, kValueTypeForSeek),
IKey("g", kMaxSequenceNumber, kValueTypeForSeek), Shorten(IKey("foo", 100, kTypeValue),
Shorten(IKey("foo", 100, kTypeValue), IKey("hello", 200, kTypeValue))); IKey("hello", 200, kTypeValue)));
// When start user key is prefix of limit user key // When start user key is prefix of limit user key
ASSERT_EQ( ASSERT_EQ(IKey("foo", 100, kTypeValue),
IKey("foo", 100, kTypeValue), Shorten(IKey("foo", 100, kTypeValue),
Shorten(IKey("foo", 100, kTypeValue), IKey("foobar", 200, kTypeValue))); IKey("foobar", 200, kTypeValue)));
// When limit user key is prefix of start user key // When limit user key is prefix of start user key
ASSERT_EQ( ASSERT_EQ(IKey("foobar", 100, kTypeValue),
IKey("foobar", 100, kTypeValue), Shorten(IKey("foobar", 100, kTypeValue),
Shorten(IKey("foobar", 100, kTypeValue), IKey("foo", 200, kTypeValue))); IKey("foo", 200, kTypeValue)));
} }
TEST(FormatTest, InternalKeyShortestSuccessor) { TEST(FormatTest, InternalKeyShortestSuccessor) {
@ -111,18 +105,8 @@ TEST(FormatTest, InternalKeyShortestSuccessor) {
ShortSuccessor(IKey("\xff\xff", 100, kTypeValue))); ShortSuccessor(IKey("\xff\xff", 100, kTypeValue)));
} }
TEST(FormatTest, ParsedInternalKeyDebugString) {
ParsedInternalKey key("The \"key\" in 'single quotes'", 42, kTypeValue);
ASSERT_EQ("'The \"key\" in 'single quotes'' @ 42 : 1", key.DebugString());
}
TEST(FormatTest, InternalKeyDebugString) {
InternalKey key("The \"key\" in 'single quotes'", 42, kTypeValue);
ASSERT_EQ("'The \"key\" in 'single quotes'' @ 42 : 1", key.DebugString());
InternalKey invalid_key;
ASSERT_EQ("(bad)", invalid_key.DebugString());
}
} // namespace leveldb } // namespace leveldb
int main(int argc, char** argv) {
return leveldb::test::RunAllTests();
}

View File

@ -2,10 +2,7 @@
// Use of this source code is governed by a BSD-style license that can be // Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file. See the AUTHORS file for names of contributors. // found in the LICENSE file. See the AUTHORS file for names of contributors.
#include "leveldb/dumpfile.h" #include <stdio.h>
#include <cstdio>
#include "db/dbformat.h" #include "db/dbformat.h"
#include "db/filename.h" #include "db/filename.h"
#include "db/log_reader.h" #include "db/log_reader.h"
@ -38,7 +35,8 @@ bool GuessType(const std::string& fname, FileType* type) {
// Notified when log reader encounters corruption. // Notified when log reader encounters corruption.
class CorruptionReporter : public log::Reader::Reporter { class CorruptionReporter : public log::Reader::Reporter {
public: public:
void Corruption(size_t bytes, const Status& status) override { WritableFile* dst_;
virtual void Corruption(size_t bytes, const Status& status) {
std::string r = "corruption: "; std::string r = "corruption: ";
AppendNumberTo(&r, bytes); AppendNumberTo(&r, bytes);
r += " bytes; "; r += " bytes; ";
@ -46,8 +44,6 @@ class CorruptionReporter : public log::Reader::Reporter {
r.push_back('\n'); r.push_back('\n');
dst_->Append(r); dst_->Append(r);
} }
WritableFile* dst_;
}; };
// Print contents of a log file. (*func)() is called on every record. // Print contents of a log file. (*func)() is called on every record.
@ -74,7 +70,8 @@ Status PrintLogContents(Env* env, const std::string& fname,
// Called on every item found in a WriteBatch. // Called on every item found in a WriteBatch.
class WriteBatchItemPrinter : public WriteBatch::Handler { class WriteBatchItemPrinter : public WriteBatch::Handler {
public: public:
void Put(const Slice& key, const Slice& value) override { WritableFile* dst_;
virtual void Put(const Slice& key, const Slice& value) {
std::string r = " put '"; std::string r = " put '";
AppendEscapedStringTo(&r, key); AppendEscapedStringTo(&r, key);
r += "' '"; r += "' '";
@ -82,16 +79,15 @@ class WriteBatchItemPrinter : public WriteBatch::Handler {
r += "'\n"; r += "'\n";
dst_->Append(r); dst_->Append(r);
} }
void Delete(const Slice& key) override { virtual void Delete(const Slice& key) {
std::string r = " del '"; std::string r = " del '";
AppendEscapedStringTo(&r, key); AppendEscapedStringTo(&r, key);
r += "'\n"; r += "'\n";
dst_->Append(r); dst_->Append(r);
} }
WritableFile* dst_;
}; };
// Called on every log record (each one of which is a WriteBatch) // Called on every log record (each one of which is a WriteBatch)
// found in a kLogFile. // found in a kLogFile.
static void WriteBatchPrinter(uint64_t pos, Slice record, WritableFile* dst) { static void WriteBatchPrinter(uint64_t pos, Slice record, WritableFile* dst) {
@ -146,8 +142,8 @@ Status DumpDescriptor(Env* env, const std::string& fname, WritableFile* dst) {
Status DumpTable(Env* env, const std::string& fname, WritableFile* dst) { Status DumpTable(Env* env, const std::string& fname, WritableFile* dst) {
uint64_t file_size; uint64_t file_size;
RandomAccessFile* file = nullptr; RandomAccessFile* file = NULL;
Table* table = nullptr; Table* table = NULL;
Status s = env->GetFileSize(fname, &file_size); Status s = env->GetFileSize(fname, &file_size);
if (s.ok()) { if (s.ok()) {
s = env->NewRandomAccessFile(fname, &file); s = env->NewRandomAccessFile(fname, &file);
@ -217,12 +213,9 @@ Status DumpFile(Env* env, const std::string& fname, WritableFile* dst) {
return Status::InvalidArgument(fname + ": unknown file type"); return Status::InvalidArgument(fname + ": unknown file type");
} }
switch (ftype) { switch (ftype) {
case kLogFile: case kLogFile: return DumpLog(env, fname, dst);
return DumpLog(env, fname, dst); case kDescriptorFile: return DumpDescriptor(env, fname, dst);
case kDescriptorFile: case kTableFile: return DumpTable(env, fname, dst);
return DumpDescriptor(env, fname, dst);
case kTableFile:
return DumpTable(env, fname, dst);
default: default:
break; break;
} }

View File

@ -6,23 +6,21 @@
// the last "sync". It then checks for data loss errors by purposely dropping // the last "sync". It then checks for data loss errors by purposely dropping
// file data (or entire files) not protected by a "sync". // file data (or entire files) not protected by a "sync".
#include "leveldb/db.h"
#include <map> #include <map>
#include <set> #include <set>
#include "gtest/gtest.h"
#include "db/db_impl.h" #include "db/db_impl.h"
#include "db/filename.h" #include "db/filename.h"
#include "db/log_format.h" #include "db/log_format.h"
#include "db/version_set.h" #include "db/version_set.h"
#include "leveldb/cache.h" #include "leveldb/cache.h"
#include "leveldb/db.h"
#include "leveldb/env.h" #include "leveldb/env.h"
#include "leveldb/table.h" #include "leveldb/table.h"
#include "leveldb/write_batch.h" #include "leveldb/write_batch.h"
#include "port/port.h"
#include "port/thread_annotations.h"
#include "util/logging.h" #include "util/logging.h"
#include "util/mutexlock.h" #include "util/mutexlock.h"
#include "util/testharness.h"
#include "util/testutil.h" #include "util/testutil.h"
namespace leveldb { namespace leveldb {
@ -36,7 +34,7 @@ class FaultInjectionTestEnv;
namespace { namespace {
// Assume a filename, and not a directory name like "/foo/bar/" // Assume a filename, and not a directory name like "/foo/bar/"
static std::string GetDirName(const std::string& filename) { static std::string GetDirName(const std::string filename) {
size_t found = filename.find_last_of("/\\"); size_t found = filename.find_last_of("/\\");
if (found == std::string::npos) { if (found == std::string::npos) {
return ""; return "";
@ -56,7 +54,8 @@ Status Truncate(const std::string& filename, uint64_t length) {
SequentialFile* orig_file; SequentialFile* orig_file;
Status s = env->NewSequentialFile(filename, &orig_file); Status s = env->NewSequentialFile(filename, &orig_file);
if (!s.ok()) return s; if (!s.ok())
return s;
char* scratch = new char[length]; char* scratch = new char[length];
leveldb::Slice result; leveldb::Slice result;
@ -72,7 +71,7 @@ Status Truncate(const std::string& filename, uint64_t length) {
if (s.ok()) { if (s.ok()) {
s = env->RenameFile(tmp_name, filename); s = env->RenameFile(tmp_name, filename);
} else { } else {
env->RemoveFile(tmp_name); env->DeleteFile(tmp_name);
} }
} }
} }
@ -84,15 +83,15 @@ Status Truncate(const std::string& filename, uint64_t length) {
struct FileState { struct FileState {
std::string filename_; std::string filename_;
int64_t pos_; ssize_t pos_;
int64_t pos_at_last_sync_; ssize_t pos_at_last_sync_;
int64_t pos_at_last_flush_; ssize_t pos_at_last_flush_;
FileState(const std::string& filename) FileState(const std::string& filename)
: filename_(filename), : filename_(filename),
pos_(-1), pos_(-1),
pos_at_last_sync_(-1), pos_at_last_sync_(-1),
pos_at_last_flush_(-1) {} pos_at_last_flush_(-1) { }
FileState() : pos_(-1), pos_at_last_sync_(-1), pos_at_last_flush_(-1) {} FileState() : pos_(-1), pos_at_last_sync_(-1), pos_at_last_flush_(-1) {}
@ -107,13 +106,14 @@ struct FileState {
// is written to or sync'ed. // is written to or sync'ed.
class TestWritableFile : public WritableFile { class TestWritableFile : public WritableFile {
public: public:
TestWritableFile(const FileState& state, WritableFile* f, TestWritableFile(const FileState& state,
WritableFile* f,
FaultInjectionTestEnv* env); FaultInjectionTestEnv* env);
~TestWritableFile() override; virtual ~TestWritableFile();
Status Append(const Slice& data) override; virtual Status Append(const Slice& data);
Status Close() override; virtual Status Close();
Status Flush() override; virtual Status Flush();
Status Sync() override; virtual Status Sync();
private: private:
FileState state_; FileState state_;
@ -126,19 +126,18 @@ class TestWritableFile : public WritableFile {
class FaultInjectionTestEnv : public EnvWrapper { class FaultInjectionTestEnv : public EnvWrapper {
public: public:
FaultInjectionTestEnv() FaultInjectionTestEnv() : EnvWrapper(Env::Default()), filesystem_active_(true) {}
: EnvWrapper(Env::Default()), filesystem_active_(true) {} virtual ~FaultInjectionTestEnv() { }
~FaultInjectionTestEnv() override = default; virtual Status NewWritableFile(const std::string& fname,
Status NewWritableFile(const std::string& fname, WritableFile** result);
WritableFile** result) override; virtual Status NewAppendableFile(const std::string& fname,
Status NewAppendableFile(const std::string& fname, WritableFile** result);
WritableFile** result) override; virtual Status DeleteFile(const std::string& f);
Status RemoveFile(const std::string& f) override; virtual Status RenameFile(const std::string& s, const std::string& t);
Status RenameFile(const std::string& s, const std::string& t) override;
void WritableFileClosed(const FileState& state); void WritableFileClosed(const FileState& state);
Status DropUnsyncedFileData(); Status DropUnsyncedFileData();
Status RemoveFilesCreatedAfterLastDirSync(); Status DeleteFilesCreatedAfterLastDirSync();
void DirWasSynced(); void DirWasSynced();
bool IsFileCreatedSinceLastDirSync(const std::string& filename); bool IsFileCreatedSinceLastDirSync(const std::string& filename);
void ResetState(); void ResetState();
@ -147,26 +146,24 @@ class FaultInjectionTestEnv : public EnvWrapper {
// system reset. Setting to inactive will freeze our saved filesystem state so // system reset. Setting to inactive will freeze our saved filesystem state so
// that it will stop being recorded. It can then be reset back to the state at // that it will stop being recorded. It can then be reset back to the state at
// the time of the reset. // the time of the reset.
bool IsFilesystemActive() LOCKS_EXCLUDED(mutex_) { bool IsFilesystemActive() const { return filesystem_active_; }
MutexLock l(&mutex_); void SetFilesystemActive(bool active) { filesystem_active_ = active; }
return filesystem_active_;
}
void SetFilesystemActive(bool active) LOCKS_EXCLUDED(mutex_) {
MutexLock l(&mutex_);
filesystem_active_ = active;
}
private: private:
port::Mutex mutex_; port::Mutex mutex_;
std::map<std::string, FileState> db_file_state_ GUARDED_BY(mutex_); std::map<std::string, FileState> db_file_state_;
std::set<std::string> new_files_since_last_dir_sync_ GUARDED_BY(mutex_); std::set<std::string> new_files_since_last_dir_sync_;
bool filesystem_active_ GUARDED_BY(mutex_); // Record flushes, syncs, writes bool filesystem_active_; // Record flushes, syncs, writes
}; };
TestWritableFile::TestWritableFile(const FileState& state, WritableFile* f, TestWritableFile::TestWritableFile(const FileState& state,
WritableFile* f,
FaultInjectionTestEnv* env) FaultInjectionTestEnv* env)
: state_(state), target_(f), writable_file_opened_(true), env_(env) { : state_(state),
assert(f != nullptr); target_(f),
writable_file_opened_(true),
env_(env) {
assert(f != NULL);
} }
TestWritableFile::~TestWritableFile() { TestWritableFile::~TestWritableFile() {
@ -268,11 +265,10 @@ Status FaultInjectionTestEnv::NewAppendableFile(const std::string& fname,
Status FaultInjectionTestEnv::DropUnsyncedFileData() { Status FaultInjectionTestEnv::DropUnsyncedFileData() {
Status s; Status s;
MutexLock l(&mutex_); MutexLock l(&mutex_);
for (const auto& kvp : db_file_state_) { for (std::map<std::string, FileState>::const_iterator it =
if (!s.ok()) { db_file_state_.begin();
break; s.ok() && it != db_file_state_.end(); ++it) {
} const FileState& state = it->second;
const FileState& state = kvp.second;
if (!state.IsFullySynced()) { if (!state.IsFullySynced()) {
s = state.DropUnsyncedData(); s = state.DropUnsyncedData();
} }
@ -298,9 +294,9 @@ void FaultInjectionTestEnv::UntrackFile(const std::string& f) {
new_files_since_last_dir_sync_.erase(f); new_files_since_last_dir_sync_.erase(f);
} }
Status FaultInjectionTestEnv::RemoveFile(const std::string& f) { Status FaultInjectionTestEnv::DeleteFile(const std::string& f) {
Status s = EnvWrapper::RemoveFile(f); Status s = EnvWrapper::DeleteFile(f);
EXPECT_LEVELDB_OK(s); ASSERT_OK(s);
if (s.ok()) { if (s.ok()) {
UntrackFile(f); UntrackFile(f);
} }
@ -332,23 +328,22 @@ void FaultInjectionTestEnv::ResetState() {
// Since we are not destroying the database, the existing files // Since we are not destroying the database, the existing files
// should keep their recorded synced/flushed state. Therefore // should keep their recorded synced/flushed state. Therefore
// we do not reset db_file_state_ and new_files_since_last_dir_sync_. // we do not reset db_file_state_ and new_files_since_last_dir_sync_.
MutexLock l(&mutex_);
SetFilesystemActive(true); SetFilesystemActive(true);
} }
Status FaultInjectionTestEnv::RemoveFilesCreatedAfterLastDirSync() { Status FaultInjectionTestEnv::DeleteFilesCreatedAfterLastDirSync() {
// Because RemoveFile access this container make a copy to avoid deadlock // Because DeleteFile access this container make a copy to avoid deadlock
mutex_.Lock(); mutex_.Lock();
std::set<std::string> new_files(new_files_since_last_dir_sync_.begin(), std::set<std::string> new_files(new_files_since_last_dir_sync_.begin(),
new_files_since_last_dir_sync_.end()); new_files_since_last_dir_sync_.end());
mutex_.Unlock(); mutex_.Unlock();
Status status; Status s;
for (const auto& new_file : new_files) { std::set<std::string>::const_iterator it;
Status remove_status = RemoveFile(new_file); for (it = new_files.begin(); s.ok() && it != new_files.end(); ++it) {
if (!remove_status.ok() && status.ok()) { s = DeleteFile(*it);
status = std::move(remove_status);
}
} }
return status; return s;
} }
void FaultInjectionTestEnv::WritableFileClosed(const FileState& state) { void FaultInjectionTestEnv::WritableFileClosed(const FileState& state) {
@ -357,11 +352,11 @@ void FaultInjectionTestEnv::WritableFileClosed(const FileState& state) {
} }
Status FileState::DropUnsyncedData() const { Status FileState::DropUnsyncedData() const {
int64_t sync_pos = pos_at_last_sync_ == -1 ? 0 : pos_at_last_sync_; ssize_t sync_pos = pos_at_last_sync_ == -1 ? 0 : pos_at_last_sync_;
return Truncate(filename_, sync_pos); return Truncate(filename_, sync_pos);
} }
class FaultInjectionTest : public testing::Test { class FaultInjectionTest {
public: public:
enum ExpectedVerifResult { VAL_EXPECT_NO_ERROR, VAL_EXPECT_ERROR }; enum ExpectedVerifResult { VAL_EXPECT_NO_ERROR, VAL_EXPECT_ERROR };
enum ResetMethod { RESET_DROP_UNSYNCED_DATA, RESET_DELETE_UNSYNCED_FILES }; enum ResetMethod { RESET_DROP_UNSYNCED_DATA, RESET_DELETE_UNSYNCED_FILES };
@ -375,8 +370,8 @@ class FaultInjectionTest : public testing::Test {
FaultInjectionTest() FaultInjectionTest()
: env_(new FaultInjectionTestEnv), : env_(new FaultInjectionTestEnv),
tiny_cache_(NewLRUCache(100)), tiny_cache_(NewLRUCache(100)),
db_(nullptr) { db_(NULL) {
dbname_ = testing::TempDir() + "fault_test"; dbname_ = test::TmpDir() + "/fault_test";
DestroyDB(dbname_, Options()); // Destroy any db from earlier run DestroyDB(dbname_, Options()); // Destroy any db from earlier run
options_.reuse_logs = true; options_.reuse_logs = true;
options_.env = env_; options_.env = env_;
@ -392,7 +387,9 @@ class FaultInjectionTest : public testing::Test {
delete env_; delete env_;
} }
void ReuseLogs(bool reuse) { options_.reuse_logs = reuse; } void ReuseLogs(bool reuse) {
options_.reuse_logs = reuse;
}
void Build(int start_idx, int num_vals) { void Build(int start_idx, int num_vals) {
std::string key_space, value_space; std::string key_space, value_space;
@ -402,7 +399,7 @@ class FaultInjectionTest : public testing::Test {
batch.Clear(); batch.Clear();
batch.Put(key, Value(i, &value_space)); batch.Put(key, Value(i, &value_space));
WriteOptions options; WriteOptions options;
ASSERT_LEVELDB_OK(db_->Write(options, &batch)); ASSERT_OK(db_->Write(options, &batch));
} }
} }
@ -424,10 +421,10 @@ class FaultInjectionTest : public testing::Test {
s = ReadValue(i, &val); s = ReadValue(i, &val);
if (expected == VAL_EXPECT_NO_ERROR) { if (expected == VAL_EXPECT_NO_ERROR) {
if (s.ok()) { if (s.ok()) {
EXPECT_EQ(value_space, val); ASSERT_EQ(value_space, val);
} }
} else if (s.ok()) { } else if (s.ok()) {
std::fprintf(stderr, "Expected an error at %d, but was OK\n", i); fprintf(stderr, "Expected an error at %d, but was OK\n", i);
s = Status::IOError(dbname_, "Expected value error:"); s = Status::IOError(dbname_, "Expected value error:");
} else { } else {
s = Status::OK(); // An expected error s = Status::OK(); // An expected error
@ -439,7 +436,7 @@ class FaultInjectionTest : public testing::Test {
// Return the ith key // Return the ith key
Slice Key(int i, std::string* storage) const { Slice Key(int i, std::string* storage) const {
char buf[100]; char buf[100];
std::snprintf(buf, sizeof(buf), "%016d", i); snprintf(buf, sizeof(buf), "%016d", i);
storage->assign(buf, strlen(buf)); storage->assign(buf, strlen(buf));
return Slice(*storage); return Slice(*storage);
} }
@ -452,20 +449,21 @@ class FaultInjectionTest : public testing::Test {
Status OpenDB() { Status OpenDB() {
delete db_; delete db_;
db_ = nullptr; db_ = NULL;
env_->ResetState(); env_->ResetState();
return DB::Open(options_, dbname_, &db_); return DB::Open(options_, dbname_, &db_);
} }
void CloseDB() { void CloseDB() {
delete db_; delete db_;
db_ = nullptr; db_ = NULL;
} }
void DeleteAllData() { void DeleteAllData() {
Iterator* iter = db_->NewIterator(ReadOptions()); Iterator* iter = db_->NewIterator(ReadOptions());
WriteOptions options;
for (iter->SeekToFirst(); iter->Valid(); iter->Next()) { for (iter->SeekToFirst(); iter->Valid(); iter->Next()) {
ASSERT_LEVELDB_OK(db_->Delete(WriteOptions(), iter->key())); ASSERT_OK(db_->Delete(WriteOptions(), iter->key()));
} }
delete iter; delete iter;
@ -474,10 +472,10 @@ class FaultInjectionTest : public testing::Test {
void ResetDBState(ResetMethod reset_method) { void ResetDBState(ResetMethod reset_method) {
switch (reset_method) { switch (reset_method) {
case RESET_DROP_UNSYNCED_DATA: case RESET_DROP_UNSYNCED_DATA:
ASSERT_LEVELDB_OK(env_->DropUnsyncedFileData()); ASSERT_OK(env_->DropUnsyncedFileData());
break; break;
case RESET_DELETE_UNSYNCED_FILES: case RESET_DELETE_UNSYNCED_FILES:
ASSERT_LEVELDB_OK(env_->RemoveFilesCreatedAfterLastDirSync()); ASSERT_OK(env_->DeleteFilesCreatedAfterLastDirSync());
break; break;
default: default:
assert(false); assert(false);
@ -487,39 +485,40 @@ class FaultInjectionTest : public testing::Test {
void PartialCompactTestPreFault(int num_pre_sync, int num_post_sync) { void PartialCompactTestPreFault(int num_pre_sync, int num_post_sync) {
DeleteAllData(); DeleteAllData();
Build(0, num_pre_sync); Build(0, num_pre_sync);
db_->CompactRange(nullptr, nullptr); db_->CompactRange(NULL, NULL);
Build(num_pre_sync, num_post_sync); Build(num_pre_sync, num_post_sync);
} }
void PartialCompactTestReopenWithFault(ResetMethod reset_method, void PartialCompactTestReopenWithFault(ResetMethod reset_method,
int num_pre_sync, int num_post_sync) { int num_pre_sync,
int num_post_sync) {
env_->SetFilesystemActive(false); env_->SetFilesystemActive(false);
CloseDB(); CloseDB();
ResetDBState(reset_method); ResetDBState(reset_method);
ASSERT_LEVELDB_OK(OpenDB()); ASSERT_OK(OpenDB());
ASSERT_LEVELDB_OK( ASSERT_OK(Verify(0, num_pre_sync, FaultInjectionTest::VAL_EXPECT_NO_ERROR));
Verify(0, num_pre_sync, FaultInjectionTest::VAL_EXPECT_NO_ERROR)); ASSERT_OK(Verify(num_pre_sync, num_post_sync, FaultInjectionTest::VAL_EXPECT_ERROR));
ASSERT_LEVELDB_OK(Verify(num_pre_sync, num_post_sync,
FaultInjectionTest::VAL_EXPECT_ERROR));
} }
void NoWriteTestPreFault() {} void NoWriteTestPreFault() {
}
void NoWriteTestReopenWithFault(ResetMethod reset_method) { void NoWriteTestReopenWithFault(ResetMethod reset_method) {
CloseDB(); CloseDB();
ResetDBState(reset_method); ResetDBState(reset_method);
ASSERT_LEVELDB_OK(OpenDB()); ASSERT_OK(OpenDB());
} }
void DoTest() { void DoTest() {
Random rnd(0); Random rnd(0);
ASSERT_LEVELDB_OK(OpenDB()); ASSERT_OK(OpenDB());
for (size_t idx = 0; idx < kNumIterations; idx++) { for (size_t idx = 0; idx < kNumIterations; idx++) {
int num_pre_sync = rnd.Uniform(kMaxNumValues); int num_pre_sync = rnd.Uniform(kMaxNumValues);
int num_post_sync = rnd.Uniform(kMaxNumValues); int num_post_sync = rnd.Uniform(kMaxNumValues);
PartialCompactTestPreFault(num_pre_sync, num_post_sync); PartialCompactTestPreFault(num_pre_sync, num_post_sync);
PartialCompactTestReopenWithFault(RESET_DROP_UNSYNCED_DATA, num_pre_sync, PartialCompactTestReopenWithFault(RESET_DROP_UNSYNCED_DATA,
num_pre_sync,
num_post_sync); num_post_sync);
NoWriteTestPreFault(); NoWriteTestPreFault();
@ -529,7 +528,8 @@ class FaultInjectionTest : public testing::Test {
// No new files created so we expect all values since no files will be // No new files created so we expect all values since no files will be
// dropped. // dropped.
PartialCompactTestReopenWithFault(RESET_DELETE_UNSYNCED_FILES, PartialCompactTestReopenWithFault(RESET_DELETE_UNSYNCED_FILES,
num_pre_sync + num_post_sync, 0); num_pre_sync + num_post_sync,
0);
NoWriteTestPreFault(); NoWriteTestPreFault();
NoWriteTestReopenWithFault(RESET_DELETE_UNSYNCED_FILES); NoWriteTestReopenWithFault(RESET_DELETE_UNSYNCED_FILES);
@ -537,14 +537,18 @@ class FaultInjectionTest : public testing::Test {
} }
}; };
TEST_F(FaultInjectionTest, FaultTestNoLogReuse) { TEST(FaultInjectionTest, FaultTestNoLogReuse) {
ReuseLogs(false); ReuseLogs(false);
DoTest(); DoTest();
} }
TEST_F(FaultInjectionTest, FaultTestWithLogReuse) { TEST(FaultInjectionTest, FaultTestWithLogReuse) {
ReuseLogs(true); ReuseLogs(true);
DoTest(); DoTest();
} }
} // namespace leveldb } // namespace leveldb
int main(int argc, char** argv) {
return leveldb::test::RunAllTests();
}

View File

@ -2,11 +2,9 @@
// Use of this source code is governed by a BSD-style license that can be // Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file. See the AUTHORS file for names of contributors. // found in the LICENSE file. See the AUTHORS file for names of contributors.
#include <ctype.h>
#include <stdio.h>
#include "db/filename.h" #include "db/filename.h"
#include <cassert>
#include <cstdio>
#include "db/dbformat.h" #include "db/dbformat.h"
#include "leveldb/env.h" #include "leveldb/env.h"
#include "util/logging.h" #include "util/logging.h"
@ -14,37 +12,38 @@
namespace leveldb { namespace leveldb {
// A utility routine: write "data" to the named file and Sync() it. // A utility routine: write "data" to the named file and Sync() it.
Status WriteStringToFileSync(Env* env, const Slice& data, extern Status WriteStringToFileSync(Env* env, const Slice& data,
const std::string& fname); const std::string& fname);
static std::string MakeFileName(const std::string& dbname, uint64_t number, static std::string MakeFileName(const std::string& name, uint64_t number,
const char* suffix) { const char* suffix) {
char buf[100]; char buf[100];
std::snprintf(buf, sizeof(buf), "/%06llu.%s", snprintf(buf, sizeof(buf), "/%06llu.%s",
static_cast<unsigned long long>(number), suffix); static_cast<unsigned long long>(number),
return dbname + buf; suffix);
return name + buf;
} }
std::string LogFileName(const std::string& dbname, uint64_t number) { std::string LogFileName(const std::string& name, uint64_t number) {
assert(number > 0); assert(number > 0);
return MakeFileName(dbname, number, "log"); return MakeFileName(name, number, "log");
} }
std::string TableFileName(const std::string& dbname, uint64_t number) { std::string TableFileName(const std::string& name, uint64_t number) {
assert(number > 0); assert(number > 0);
return MakeFileName(dbname, number, "ldb"); return MakeFileName(name, number, "ldb");
} }
std::string SSTTableFileName(const std::string& dbname, uint64_t number) { std::string SSTTableFileName(const std::string& name, uint64_t number) {
assert(number > 0); assert(number > 0);
return MakeFileName(dbname, number, "sst"); return MakeFileName(name, number, "sst");
} }
std::string DescriptorFileName(const std::string& dbname, uint64_t number) { std::string DescriptorFileName(const std::string& dbname, uint64_t number) {
assert(number > 0); assert(number > 0);
char buf[100]; char buf[100];
std::snprintf(buf, sizeof(buf), "/MANIFEST-%06llu", snprintf(buf, sizeof(buf), "/MANIFEST-%06llu",
static_cast<unsigned long long>(number)); static_cast<unsigned long long>(number));
return dbname + buf; return dbname + buf;
} }
@ -52,7 +51,9 @@ std::string CurrentFileName(const std::string& dbname) {
return dbname + "/CURRENT"; return dbname + "/CURRENT";
} }
std::string LockFileName(const std::string& dbname) { return dbname + "/LOCK"; } std::string LockFileName(const std::string& dbname) {
return dbname + "/LOCK";
}
std::string TempFileName(const std::string& dbname, uint64_t number) { std::string TempFileName(const std::string& dbname, uint64_t number) {
assert(number > 0); assert(number > 0);
@ -68,6 +69,7 @@ std::string OldInfoLogFileName(const std::string& dbname) {
return dbname + "/LOG.old"; return dbname + "/LOG.old";
} }
// Owned filenames have the form: // Owned filenames have the form:
// dbname/CURRENT // dbname/CURRENT
// dbname/LOCK // dbname/LOCK
@ -75,9 +77,10 @@ std::string OldInfoLogFileName(const std::string& dbname) {
// dbname/LOG.old // dbname/LOG.old
// dbname/MANIFEST-[0-9]+ // dbname/MANIFEST-[0-9]+
// dbname/[0-9]+.(log|sst|ldb) // dbname/[0-9]+.(log|sst|ldb)
bool ParseFileName(const std::string& filename, uint64_t* number, bool ParseFileName(const std::string& fname,
uint64_t* number,
FileType* type) { FileType* type) {
Slice rest(filename); Slice rest(fname);
if (rest == "CURRENT") { if (rest == "CURRENT") {
*number = 0; *number = 0;
*type = kCurrentFile; *type = kCurrentFile;
@ -133,7 +136,7 @@ Status SetCurrentFile(Env* env, const std::string& dbname,
s = env->RenameFile(tmp, CurrentFileName(dbname)); s = env->RenameFile(tmp, CurrentFileName(dbname));
} }
if (!s.ok()) { if (!s.ok()) {
env->RemoveFile(tmp); env->DeleteFile(tmp);
} }
return s; return s;
} }

View File

@ -7,9 +7,8 @@
#ifndef STORAGE_LEVELDB_DB_FILENAME_H_ #ifndef STORAGE_LEVELDB_DB_FILENAME_H_
#define STORAGE_LEVELDB_DB_FILENAME_H_ #define STORAGE_LEVELDB_DB_FILENAME_H_
#include <cstdint> #include <stdint.h>
#include <string> #include <string>
#include "leveldb/slice.h" #include "leveldb/slice.h"
#include "leveldb/status.h" #include "leveldb/status.h"
#include "port/port.h" #include "port/port.h"
@ -31,52 +30,55 @@ enum FileType {
// Return the name of the log file with the specified number // Return the name of the log file with the specified number
// in the db named by "dbname". The result will be prefixed with // in the db named by "dbname". The result will be prefixed with
// "dbname". // "dbname".
std::string LogFileName(const std::string& dbname, uint64_t number); extern std::string LogFileName(const std::string& dbname, uint64_t number);
// Return the name of the sstable with the specified number // Return the name of the sstable with the specified number
// in the db named by "dbname". The result will be prefixed with // in the db named by "dbname". The result will be prefixed with
// "dbname". // "dbname".
std::string TableFileName(const std::string& dbname, uint64_t number); extern std::string TableFileName(const std::string& dbname, uint64_t number);
// Return the legacy file name for an sstable with the specified number // Return the legacy file name for an sstable with the specified number
// in the db named by "dbname". The result will be prefixed with // in the db named by "dbname". The result will be prefixed with
// "dbname". // "dbname".
std::string SSTTableFileName(const std::string& dbname, uint64_t number); extern std::string SSTTableFileName(const std::string& dbname, uint64_t number);
// Return the name of the descriptor file for the db named by // Return the name of the descriptor file for the db named by
// "dbname" and the specified incarnation number. The result will be // "dbname" and the specified incarnation number. The result will be
// prefixed with "dbname". // prefixed with "dbname".
std::string DescriptorFileName(const std::string& dbname, uint64_t number); extern std::string DescriptorFileName(const std::string& dbname,
uint64_t number);
// Return the name of the current file. This file contains the name // Return the name of the current file. This file contains the name
// of the current manifest file. The result will be prefixed with // of the current manifest file. The result will be prefixed with
// "dbname". // "dbname".
std::string CurrentFileName(const std::string& dbname); extern std::string CurrentFileName(const std::string& dbname);
// Return the name of the lock file for the db named by // Return the name of the lock file for the db named by
// "dbname". The result will be prefixed with "dbname". // "dbname". The result will be prefixed with "dbname".
std::string LockFileName(const std::string& dbname); extern std::string LockFileName(const std::string& dbname);
// Return the name of a temporary file owned by the db named "dbname". // Return the name of a temporary file owned by the db named "dbname".
// The result will be prefixed with "dbname". // The result will be prefixed with "dbname".
std::string TempFileName(const std::string& dbname, uint64_t number); extern std::string TempFileName(const std::string& dbname, uint64_t number);
// Return the name of the info log file for "dbname". // Return the name of the info log file for "dbname".
std::string InfoLogFileName(const std::string& dbname); extern std::string InfoLogFileName(const std::string& dbname);
// Return the name of the old info log file for "dbname". // Return the name of the old info log file for "dbname".
std::string OldInfoLogFileName(const std::string& dbname); extern std::string OldInfoLogFileName(const std::string& dbname);
// If filename is a leveldb file, store the type of the file in *type. // If filename is a leveldb file, store the type of the file in *type.
// The number encoded in the filename is stored in *number. If the // The number encoded in the filename is stored in *number. If the
// filename was successfully parsed, returns true. Else return false. // filename was successfully parsed, returns true. Else return false.
bool ParseFileName(const std::string& filename, uint64_t* number, extern bool ParseFileName(const std::string& filename,
FileType* type); uint64_t* number,
FileType* type);
// Make the CURRENT file point to the descriptor file with the // Make the CURRENT file point to the descriptor file with the
// specified number. // specified number.
Status SetCurrentFile(Env* env, const std::string& dbname, extern Status SetCurrentFile(Env* env, const std::string& dbname,
uint64_t descriptor_number); uint64_t descriptor_number);
} // namespace leveldb } // namespace leveldb

View File

@ -4,13 +4,15 @@
#include "db/filename.h" #include "db/filename.h"
#include "gtest/gtest.h"
#include "db/dbformat.h" #include "db/dbformat.h"
#include "port/port.h" #include "port/port.h"
#include "util/logging.h" #include "util/logging.h"
#include "util/testharness.h"
namespace leveldb { namespace leveldb {
class FileNameTest { };
TEST(FileNameTest, Parse) { TEST(FileNameTest, Parse) {
Slice db; Slice db;
FileType type; FileType type;
@ -22,17 +24,17 @@ TEST(FileNameTest, Parse) {
uint64_t number; uint64_t number;
FileType type; FileType type;
} cases[] = { } cases[] = {
{"100.log", 100, kLogFile}, { "100.log", 100, kLogFile },
{"0.log", 0, kLogFile}, { "0.log", 0, kLogFile },
{"0.sst", 0, kTableFile}, { "0.sst", 0, kTableFile },
{"0.ldb", 0, kTableFile}, { "0.ldb", 0, kTableFile },
{"CURRENT", 0, kCurrentFile}, { "CURRENT", 0, kCurrentFile },
{"LOCK", 0, kDBLockFile}, { "LOCK", 0, kDBLockFile },
{"MANIFEST-2", 2, kDescriptorFile}, { "MANIFEST-2", 2, kDescriptorFile },
{"MANIFEST-7", 7, kDescriptorFile}, { "MANIFEST-7", 7, kDescriptorFile },
{"LOG", 0, kInfoLogFile}, { "LOG", 0, kInfoLogFile },
{"LOG.old", 0, kInfoLogFile}, { "LOG.old", 0, kInfoLogFile },
{"18446744073709551615.log", 18446744073709551615ull, kLogFile}, { "18446744073709551615.log", 18446744073709551615ull, kLogFile },
}; };
for (int i = 0; i < sizeof(cases) / sizeof(cases[0]); i++) { for (int i = 0; i < sizeof(cases) / sizeof(cases[0]); i++) {
std::string f = cases[i].fname; std::string f = cases[i].fname;
@ -42,28 +44,30 @@ TEST(FileNameTest, Parse) {
} }
// Errors // Errors
static const char* errors[] = {"", static const char* errors[] = {
"foo", "",
"foo-dx-100.log", "foo",
".log", "foo-dx-100.log",
"", ".log",
"manifest", "",
"CURREN", "manifest",
"CURRENTX", "CURREN",
"MANIFES", "CURRENTX",
"MANIFEST", "MANIFES",
"MANIFEST-", "MANIFEST",
"XMANIFEST-3", "MANIFEST-",
"MANIFEST-3x", "XMANIFEST-3",
"LOC", "MANIFEST-3x",
"LOCKx", "LOC",
"LO", "LOCKx",
"LOGx", "LO",
"18446744073709551616.log", "LOGx",
"184467440737095516150.log", "18446744073709551616.log",
"100", "184467440737095516150.log",
"100.", "100",
"100.lop"}; "100.",
"100.lop"
};
for (int i = 0; i < sizeof(errors) / sizeof(errors[0]); i++) { for (int i = 0; i < sizeof(errors) / sizeof(errors[0]); i++) {
std::string f = errors[i]; std::string f = errors[i];
ASSERT_TRUE(!ParseFileName(f, &number, &type)) << f; ASSERT_TRUE(!ParseFileName(f, &number, &type)) << f;
@ -110,18 +114,10 @@ TEST(FileNameTest, Construction) {
ASSERT_TRUE(ParseFileName(fname.c_str() + 4, &number, &type)); ASSERT_TRUE(ParseFileName(fname.c_str() + 4, &number, &type));
ASSERT_EQ(999, number); ASSERT_EQ(999, number);
ASSERT_EQ(kTempFile, type); ASSERT_EQ(kTempFile, type);
fname = InfoLogFileName("foo");
ASSERT_EQ("foo/", std::string(fname.data(), 4));
ASSERT_TRUE(ParseFileName(fname.c_str() + 4, &number, &type));
ASSERT_EQ(0, number);
ASSERT_EQ(kInfoLogFile, type);
fname = OldInfoLogFileName("foo");
ASSERT_EQ("foo/", std::string(fname.data(), 4));
ASSERT_TRUE(ParseFileName(fname.c_str() + 4, &number, &type));
ASSERT_EQ(0, number);
ASSERT_EQ(kInfoLogFile, type);
} }
} // namespace leveldb } // namespace leveldb
int main(int argc, char** argv) {
return leveldb::test::RunAllTests();
}

View File

@ -2,8 +2,7 @@
// Use of this source code is governed by a BSD-style license that can be // Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file. See the AUTHORS file for names of contributors. // found in the LICENSE file. See the AUTHORS file for names of contributors.
#include <cstdio> #include <stdio.h>
#include "leveldb/dumpfile.h" #include "leveldb/dumpfile.h"
#include "leveldb/env.h" #include "leveldb/env.h"
#include "leveldb/status.h" #include "leveldb/status.h"
@ -13,13 +12,13 @@ namespace {
class StdoutPrinter : public WritableFile { class StdoutPrinter : public WritableFile {
public: public:
Status Append(const Slice& data) override { virtual Status Append(const Slice& data) {
fwrite(data.data(), 1, data.size(), stdout); fwrite(data.data(), 1, data.size(), stdout);
return Status::OK(); return Status::OK();
} }
Status Close() override { return Status::OK(); } virtual Status Close() { return Status::OK(); }
Status Flush() override { return Status::OK(); } virtual Status Flush() { return Status::OK(); }
Status Sync() override { return Status::OK(); } virtual Status Sync() { return Status::OK(); }
}; };
bool HandleDumpCommand(Env* env, char** files, int num) { bool HandleDumpCommand(Env* env, char** files, int num) {
@ -28,7 +27,7 @@ bool HandleDumpCommand(Env* env, char** files, int num) {
for (int i = 0; i < num; i++) { for (int i = 0; i < num; i++) {
Status s = DumpFile(env, files[i], &printer); Status s = DumpFile(env, files[i], &printer);
if (!s.ok()) { if (!s.ok()) {
std::fprintf(stderr, "%s\n", s.ToString().c_str()); fprintf(stderr, "%s\n", s.ToString().c_str());
ok = false; ok = false;
} }
} }
@ -39,10 +38,11 @@ bool HandleDumpCommand(Env* env, char** files, int num) {
} // namespace leveldb } // namespace leveldb
static void Usage() { static void Usage() {
std::fprintf( fprintf(
stderr, stderr,
"Usage: leveldbutil command...\n" "Usage: leveldbutil command...\n"
" dump files... -- dump contents of specified files\n"); " dump files... -- dump contents of specified files\n"
);
} }
int main(int argc, char** argv) { int main(int argc, char** argv) {
@ -54,7 +54,7 @@ int main(int argc, char** argv) {
} else { } else {
std::string command = argv[1]; std::string command = argv[1];
if (command == "dump") { if (command == "dump") {
ok = leveldb::HandleDumpCommand(env, argv + 2, argc - 2); ok = leveldb::HandleDumpCommand(env, argv+2, argc-2);
} else { } else {
Usage(); Usage();
ok = false; ok = false;

View File

@ -3,7 +3,7 @@
// found in the LICENSE file. See the AUTHORS file for names of contributors. // found in the LICENSE file. See the AUTHORS file for names of contributors.
// //
// Log format information shared by reader and writer. // Log format information shared by reader and writer.
// See ../doc/log_format.md for more detail. // See ../doc/log_format.txt for more detail.
#ifndef STORAGE_LEVELDB_DB_LOG_FORMAT_H_ #ifndef STORAGE_LEVELDB_DB_LOG_FORMAT_H_
#define STORAGE_LEVELDB_DB_LOG_FORMAT_H_ #define STORAGE_LEVELDB_DB_LOG_FORMAT_H_

View File

@ -4,8 +4,7 @@
#include "db/log_reader.h" #include "db/log_reader.h"
#include <cstdio> #include <stdio.h>
#include "leveldb/env.h" #include "leveldb/env.h"
#include "util/coding.h" #include "util/coding.h"
#include "util/crc32c.h" #include "util/crc32c.h"
@ -13,7 +12,8 @@
namespace leveldb { namespace leveldb {
namespace log { namespace log {
Reader::Reporter::~Reporter() = default; Reader::Reporter::~Reporter() {
}
Reader::Reader(SequentialFile* file, Reporter* reporter, bool checksum, Reader::Reader(SequentialFile* file, Reporter* reporter, bool checksum,
uint64_t initial_offset) uint64_t initial_offset)
@ -25,17 +25,20 @@ Reader::Reader(SequentialFile* file, Reporter* reporter, bool checksum,
eof_(false), eof_(false),
last_record_offset_(0), last_record_offset_(0),
end_of_buffer_offset_(0), end_of_buffer_offset_(0),
initial_offset_(initial_offset), initial_offset_(initial_offset) {
resyncing_(initial_offset > 0) {} }
Reader::~Reader() { delete[] backing_store_; } Reader::~Reader() {
delete[] backing_store_;
}
bool Reader::SkipToInitialBlock() { bool Reader::SkipToInitialBlock() {
const size_t offset_in_block = initial_offset_ % kBlockSize; size_t offset_in_block = initial_offset_ % kBlockSize;
uint64_t block_start_location = initial_offset_ - offset_in_block; uint64_t block_start_location = initial_offset_ - offset_in_block;
// Don't search a block if we'd be in the trailer // Don't search a block if we'd be in the trailer
if (offset_in_block > kBlockSize - 6) { if (offset_in_block > kBlockSize - 6) {
offset_in_block = 0;
block_start_location += kBlockSize; block_start_location += kBlockSize;
} }
@ -69,25 +72,8 @@ bool Reader::ReadRecord(Slice* record, std::string* scratch) {
Slice fragment; Slice fragment;
while (true) { while (true) {
uint64_t physical_record_offset = end_of_buffer_offset_ - buffer_.size();
const unsigned int record_type = ReadPhysicalRecord(&fragment); const unsigned int record_type = ReadPhysicalRecord(&fragment);
// ReadPhysicalRecord may have only had an empty trailer remaining in its
// internal buffer. Calculate the offset of the next physical record now
// that it has returned, properly accounting for its header size.
uint64_t physical_record_offset =
end_of_buffer_offset_ - buffer_.size() - kHeaderSize - fragment.size();
if (resyncing_) {
if (record_type == kMiddleType) {
continue;
} else if (record_type == kLastType) {
resyncing_ = false;
continue;
} else {
resyncing_ = false;
}
}
switch (record_type) { switch (record_type) {
case kFullType: case kFullType:
if (in_fragmented_record) { if (in_fragmented_record) {
@ -95,7 +81,9 @@ bool Reader::ReadRecord(Slice* record, std::string* scratch) {
// it could emit an empty kFirstType record at the tail end // it could emit an empty kFirstType record at the tail end
// of a block followed by a kFullType or kFirstType record // of a block followed by a kFullType or kFirstType record
// at the beginning of the next block. // at the beginning of the next block.
if (!scratch->empty()) { if (scratch->empty()) {
in_fragmented_record = false;
} else {
ReportCorruption(scratch->size(), "partial record without end(1)"); ReportCorruption(scratch->size(), "partial record without end(1)");
} }
} }
@ -111,7 +99,9 @@ bool Reader::ReadRecord(Slice* record, std::string* scratch) {
// it could emit an empty kFirstType record at the tail end // it could emit an empty kFirstType record at the tail end
// of a block followed by a kFullType or kFirstType record // of a block followed by a kFullType or kFirstType record
// at the beginning of the next block. // at the beginning of the next block.
if (!scratch->empty()) { if (scratch->empty()) {
in_fragmented_record = false;
} else {
ReportCorruption(scratch->size(), "partial record without end(2)"); ReportCorruption(scratch->size(), "partial record without end(2)");
} }
} }
@ -160,7 +150,7 @@ bool Reader::ReadRecord(Slice* record, std::string* scratch) {
default: { default: {
char buf[40]; char buf[40];
std::snprintf(buf, sizeof(buf), "unknown record type %u", record_type); snprintf(buf, sizeof(buf), "unknown record type %u", record_type);
ReportCorruption( ReportCorruption(
(fragment.size() + (in_fragmented_record ? scratch->size() : 0)), (fragment.size() + (in_fragmented_record ? scratch->size() : 0)),
buf); buf);
@ -173,14 +163,16 @@ bool Reader::ReadRecord(Slice* record, std::string* scratch) {
return false; return false;
} }
uint64_t Reader::LastRecordOffset() { return last_record_offset_; } uint64_t Reader::LastRecordOffset() {
return last_record_offset_;
}
void Reader::ReportCorruption(uint64_t bytes, const char* reason) { void Reader::ReportCorruption(uint64_t bytes, const char* reason) {
ReportDrop(bytes, Status::Corruption(reason)); ReportDrop(bytes, Status::Corruption(reason));
} }
void Reader::ReportDrop(uint64_t bytes, const Status& reason) { void Reader::ReportDrop(uint64_t bytes, const Status& reason) {
if (reporter_ != nullptr && if (reporter_ != NULL &&
end_of_buffer_offset_ - buffer_.size() - bytes >= initial_offset_) { end_of_buffer_offset_ - buffer_.size() - bytes >= initial_offset_) {
reporter_->Corruption(static_cast<size_t>(bytes), reason); reporter_->Corruption(static_cast<size_t>(bytes), reason);
} }

View File

@ -5,7 +5,7 @@
#ifndef STORAGE_LEVELDB_DB_LOG_READER_H_ #ifndef STORAGE_LEVELDB_DB_LOG_READER_H_
#define STORAGE_LEVELDB_DB_LOG_READER_H_ #define STORAGE_LEVELDB_DB_LOG_READER_H_
#include <cstdint> #include <stdint.h>
#include "db/log_format.h" #include "db/log_format.h"
#include "leveldb/slice.h" #include "leveldb/slice.h"
@ -24,7 +24,7 @@ class Reader {
public: public:
virtual ~Reporter(); virtual ~Reporter();
// Some corruption was detected. "bytes" is the approximate number // Some corruption was detected. "size" is the approximate number
// of bytes dropped due to the corruption. // of bytes dropped due to the corruption.
virtual void Corruption(size_t bytes, const Status& status) = 0; virtual void Corruption(size_t bytes, const Status& status) = 0;
}; };
@ -32,7 +32,7 @@ class Reader {
// Create a reader that will return log records from "*file". // Create a reader that will return log records from "*file".
// "*file" must remain live while this Reader is in use. // "*file" must remain live while this Reader is in use.
// //
// If "reporter" is non-null, it is notified whenever some data is // If "reporter" is non-NULL, it is notified whenever some data is
// dropped due to a detected corruption. "*reporter" must remain // dropped due to a detected corruption. "*reporter" must remain
// live while this Reader is in use. // live while this Reader is in use.
// //
@ -43,9 +43,6 @@ class Reader {
Reader(SequentialFile* file, Reporter* reporter, bool checksum, Reader(SequentialFile* file, Reporter* reporter, bool checksum,
uint64_t initial_offset); uint64_t initial_offset);
Reader(const Reader&) = delete;
Reader& operator=(const Reader&) = delete;
~Reader(); ~Reader();
// Read the next record into *record. Returns true if read // Read the next record into *record. Returns true if read
@ -61,6 +58,21 @@ class Reader {
uint64_t LastRecordOffset(); uint64_t LastRecordOffset();
private: private:
SequentialFile* const file_;
Reporter* const reporter_;
bool const checksum_;
char* const backing_store_;
Slice buffer_;
bool eof_; // Last Read() indicated EOF by returning < kBlockSize
// Offset of the last record returned by ReadRecord.
uint64_t last_record_offset_;
// Offset of the first location past the end of buffer_.
uint64_t end_of_buffer_offset_;
// Offset at which to start looking for the first record to return
uint64_t const initial_offset_;
// Extend record types with the following special values // Extend record types with the following special values
enum { enum {
kEof = kMaxRecordType + 1, kEof = kMaxRecordType + 1,
@ -85,25 +97,9 @@ class Reader {
void ReportCorruption(uint64_t bytes, const char* reason); void ReportCorruption(uint64_t bytes, const char* reason);
void ReportDrop(uint64_t bytes, const Status& reason); void ReportDrop(uint64_t bytes, const Status& reason);
SequentialFile* const file_; // No copying allowed
Reporter* const reporter_; Reader(const Reader&);
bool const checksum_; void operator=(const Reader&);
char* const backing_store_;
Slice buffer_;
bool eof_; // Last Read() indicated EOF by returning < kBlockSize
// Offset of the last record returned by ReadRecord.
uint64_t last_record_offset_;
// Offset of the first location past the end of buffer_.
uint64_t end_of_buffer_offset_;
// Offset at which to start looking for the first record to return
uint64_t const initial_offset_;
// True if we are resynchronizing after a seek (initial_offset_ > 0). In
// particular, a run of kMiddleType and kLastType records can be silently
// skipped in this mode
bool resyncing_;
}; };
} // namespace log } // namespace log

View File

@ -2,13 +2,13 @@
// Use of this source code is governed by a BSD-style license that can be // Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file. See the AUTHORS file for names of contributors. // found in the LICENSE file. See the AUTHORS file for names of contributors.
#include "gtest/gtest.h"
#include "db/log_reader.h" #include "db/log_reader.h"
#include "db/log_writer.h" #include "db/log_writer.h"
#include "leveldb/env.h" #include "leveldb/env.h"
#include "util/coding.h" #include "util/coding.h"
#include "util/crc32c.h" #include "util/crc32c.h"
#include "util/random.h" #include "util/random.h"
#include "util/testharness.h"
namespace leveldb { namespace leveldb {
namespace log { namespace log {
@ -27,7 +27,7 @@ static std::string BigString(const std::string& partial_string, size_t n) {
// Construct a string from a number // Construct a string from a number
static std::string NumberString(int n) { static std::string NumberString(int n) {
char buf[50]; char buf[50];
std::snprintf(buf, sizeof(buf), "%d.", n); snprintf(buf, sizeof(buf), "%d.", n);
return std::string(buf); return std::string(buf);
} }
@ -36,17 +36,90 @@ static std::string RandomSkewedString(int i, Random* rnd) {
return BigString(NumberString(i), rnd->Skewed(17)); return BigString(NumberString(i), rnd->Skewed(17));
} }
class LogTest : public testing::Test { class LogTest {
private:
class StringDest : public WritableFile {
public:
std::string contents_;
virtual Status Close() { return Status::OK(); }
virtual Status Flush() { return Status::OK(); }
virtual Status Sync() { return Status::OK(); }
virtual Status Append(const Slice& slice) {
contents_.append(slice.data(), slice.size());
return Status::OK();
}
};
class StringSource : public SequentialFile {
public:
Slice contents_;
bool force_error_;
bool returned_partial_;
StringSource() : force_error_(false), returned_partial_(false) { }
virtual Status Read(size_t n, Slice* result, char* scratch) {
ASSERT_TRUE(!returned_partial_) << "must not Read() after eof/error";
if (force_error_) {
force_error_ = false;
returned_partial_ = true;
return Status::Corruption("read error");
}
if (contents_.size() < n) {
n = contents_.size();
returned_partial_ = true;
}
*result = Slice(contents_.data(), n);
contents_.remove_prefix(n);
return Status::OK();
}
virtual Status Skip(uint64_t n) {
if (n > contents_.size()) {
contents_.clear();
return Status::NotFound("in-memory file skipepd past end");
}
contents_.remove_prefix(n);
return Status::OK();
}
};
class ReportCollector : public Reader::Reporter {
public:
size_t dropped_bytes_;
std::string message_;
ReportCollector() : dropped_bytes_(0) { }
virtual void Corruption(size_t bytes, const Status& status) {
dropped_bytes_ += bytes;
message_.append(status.ToString());
}
};
StringDest dest_;
StringSource source_;
ReportCollector report_;
bool reading_;
Writer* writer_;
Reader reader_;
// Record metadata for testing initial offset functionality
static size_t initial_offset_record_sizes_[];
static uint64_t initial_offset_last_record_offsets_[];
public: public:
LogTest() LogTest() : reading_(false),
: reading_(false), writer_(new Writer(&dest_)),
writer_(new Writer(&dest_)), reader_(&source_, &report_, true/*checksum*/,
reader_(new Reader(&source_, &report_, true /*checksum*/, 0/*initial_offset*/) {
0 /*initial_offset*/)) {} }
~LogTest() { ~LogTest() {
delete writer_; delete writer_;
delete reader_;
} }
void ReopenForAppend() { void ReopenForAppend() {
@ -59,7 +132,9 @@ class LogTest : public testing::Test {
writer_->AddRecord(Slice(msg)); writer_->AddRecord(Slice(msg));
} }
size_t WrittenBytes() const { return dest_.contents_.size(); } size_t WrittenBytes() const {
return dest_.contents_.size();
}
std::string Read() { std::string Read() {
if (!reading_) { if (!reading_) {
@ -68,7 +143,7 @@ class LogTest : public testing::Test {
} }
std::string scratch; std::string scratch;
Slice record; Slice record;
if (reader_->ReadRecord(&record, &scratch)) { if (reader_.ReadRecord(&record, &scratch)) {
return record.ToString(); return record.ToString();
} else { } else {
return "EOF"; return "EOF";
@ -89,16 +164,22 @@ class LogTest : public testing::Test {
void FixChecksum(int header_offset, int len) { void FixChecksum(int header_offset, int len) {
// Compute crc of type/len/data // Compute crc of type/len/data
uint32_t crc = crc32c::Value(&dest_.contents_[header_offset + 6], 1 + len); uint32_t crc = crc32c::Value(&dest_.contents_[header_offset+6], 1 + len);
crc = crc32c::Mask(crc); crc = crc32c::Mask(crc);
EncodeFixed32(&dest_.contents_[header_offset], crc); EncodeFixed32(&dest_.contents_[header_offset], crc);
} }
void ForceError() { source_.force_error_ = true; } void ForceError() {
source_.force_error_ = true;
}
size_t DroppedBytes() const { return report_.dropped_bytes_; } size_t DroppedBytes() const {
return report_.dropped_bytes_;
}
std::string ReportMessage() const { return report_.message_; } std::string ReportMessage() const {
return report_.message_;
}
// Returns OK iff recorded error message contains "msg" // Returns OK iff recorded error message contains "msg"
std::string MatchError(const std::string& msg) const { std::string MatchError(const std::string& msg) const {
@ -110,23 +191,18 @@ class LogTest : public testing::Test {
} }
void WriteInitialOffsetLog() { void WriteInitialOffsetLog() {
for (int i = 0; i < num_initial_offset_records_; i++) { for (int i = 0; i < 4; i++) {
std::string record(initial_offset_record_sizes_[i], std::string record(initial_offset_record_sizes_[i],
static_cast<char>('a' + i)); static_cast<char>('a' + i));
Write(record); Write(record);
} }
} }
void StartReadingAt(uint64_t initial_offset) {
delete reader_;
reader_ = new Reader(&source_, &report_, true /*checksum*/, initial_offset);
}
void CheckOffsetPastEndReturnsNoRecords(uint64_t offset_past_end) { void CheckOffsetPastEndReturnsNoRecords(uint64_t offset_past_end) {
WriteInitialOffsetLog(); WriteInitialOffsetLog();
reading_ = true; reading_ = true;
source_.contents_ = Slice(dest_.contents_); source_.contents_ = Slice(dest_.contents_);
Reader* offset_reader = new Reader(&source_, &report_, true /*checksum*/, Reader* offset_reader = new Reader(&source_, &report_, true/*checksum*/,
WrittenBytes() + offset_past_end); WrittenBytes() + offset_past_end);
Slice record; Slice record;
std::string scratch; std::string scratch;
@ -139,128 +215,40 @@ class LogTest : public testing::Test {
WriteInitialOffsetLog(); WriteInitialOffsetLog();
reading_ = true; reading_ = true;
source_.contents_ = Slice(dest_.contents_); source_.contents_ = Slice(dest_.contents_);
Reader* offset_reader = Reader* offset_reader = new Reader(&source_, &report_, true/*checksum*/,
new Reader(&source_, &report_, true /*checksum*/, initial_offset); initial_offset);
Slice record;
// Read all records from expected_record_offset through the last one. std::string scratch;
ASSERT_LT(expected_record_offset, num_initial_offset_records_); ASSERT_TRUE(offset_reader->ReadRecord(&record, &scratch));
for (; expected_record_offset < num_initial_offset_records_; ASSERT_EQ(initial_offset_record_sizes_[expected_record_offset],
++expected_record_offset) { record.size());
Slice record; ASSERT_EQ(initial_offset_last_record_offsets_[expected_record_offset],
std::string scratch; offset_reader->LastRecordOffset());
ASSERT_TRUE(offset_reader->ReadRecord(&record, &scratch)); ASSERT_EQ((char)('a' + expected_record_offset), record.data()[0]);
ASSERT_EQ(initial_offset_record_sizes_[expected_record_offset],
record.size());
ASSERT_EQ(initial_offset_last_record_offsets_[expected_record_offset],
offset_reader->LastRecordOffset());
ASSERT_EQ((char)('a' + expected_record_offset), record.data()[0]);
}
delete offset_reader; delete offset_reader;
} }
private:
class StringDest : public WritableFile {
public:
Status Close() override { return Status::OK(); }
Status Flush() override { return Status::OK(); }
Status Sync() override { return Status::OK(); }
Status Append(const Slice& slice) override {
contents_.append(slice.data(), slice.size());
return Status::OK();
}
std::string contents_;
};
class StringSource : public SequentialFile {
public:
StringSource() : force_error_(false), returned_partial_(false) {}
Status Read(size_t n, Slice* result, char* scratch) override {
EXPECT_TRUE(!returned_partial_) << "must not Read() after eof/error";
if (force_error_) {
force_error_ = false;
returned_partial_ = true;
return Status::Corruption("read error");
}
if (contents_.size() < n) {
n = contents_.size();
returned_partial_ = true;
}
*result = Slice(contents_.data(), n);
contents_.remove_prefix(n);
return Status::OK();
}
Status Skip(uint64_t n) override {
if (n > contents_.size()) {
contents_.clear();
return Status::NotFound("in-memory file skipped past end");
}
contents_.remove_prefix(n);
return Status::OK();
}
Slice contents_;
bool force_error_;
bool returned_partial_;
};
class ReportCollector : public Reader::Reporter {
public:
ReportCollector() : dropped_bytes_(0) {}
void Corruption(size_t bytes, const Status& status) override {
dropped_bytes_ += bytes;
message_.append(status.ToString());
}
size_t dropped_bytes_;
std::string message_;
};
// Record metadata for testing initial offset functionality
static size_t initial_offset_record_sizes_[];
static uint64_t initial_offset_last_record_offsets_[];
static int num_initial_offset_records_;
StringDest dest_;
StringSource source_;
ReportCollector report_;
bool reading_;
Writer* writer_;
Reader* reader_;
}; };
size_t LogTest::initial_offset_record_sizes_[] = { size_t LogTest::initial_offset_record_sizes_[] =
10000, // Two sizable records in first block {10000, // Two sizable records in first block
10000, 10000,
2 * log::kBlockSize - 1000, // Span three blocks 2 * log::kBlockSize - 1000, // Span three blocks
1, 1};
13716, // Consume all but two bytes of block 3.
log::kBlockSize - kHeaderSize, // Consume the entirety of block 4.
};
uint64_t LogTest::initial_offset_last_record_offsets_[] = { uint64_t LogTest::initial_offset_last_record_offsets_[] =
0, {0,
kHeaderSize + 10000, kHeaderSize + 10000,
2 * (kHeaderSize + 10000), 2 * (kHeaderSize + 10000),
2 * (kHeaderSize + 10000) + (2 * log::kBlockSize - 1000) + 3 * kHeaderSize, 2 * (kHeaderSize + 10000) +
2 * (kHeaderSize + 10000) + (2 * log::kBlockSize - 1000) + 3 * kHeaderSize + (2 * log::kBlockSize - 1000) + 3 * kHeaderSize};
kHeaderSize + 1,
3 * log::kBlockSize,
};
// LogTest::initial_offset_last_record_offsets_ must be defined before this.
int LogTest::num_initial_offset_records_ =
sizeof(LogTest::initial_offset_last_record_offsets_) / sizeof(uint64_t);
TEST_F(LogTest, Empty) { ASSERT_EQ("EOF", Read()); } TEST(LogTest, Empty) {
ASSERT_EQ("EOF", Read());
}
TEST_F(LogTest, ReadWrite) { TEST(LogTest, ReadWrite) {
Write("foo"); Write("foo");
Write("bar"); Write("bar");
Write(""); Write("");
@ -273,7 +261,7 @@ TEST_F(LogTest, ReadWrite) {
ASSERT_EQ("EOF", Read()); // Make sure reads at eof work ASSERT_EQ("EOF", Read()); // Make sure reads at eof work
} }
TEST_F(LogTest, ManyBlocks) { TEST(LogTest, ManyBlocks) {
for (int i = 0; i < 100000; i++) { for (int i = 0; i < 100000; i++) {
Write(NumberString(i)); Write(NumberString(i));
} }
@ -283,7 +271,7 @@ TEST_F(LogTest, ManyBlocks) {
ASSERT_EQ("EOF", Read()); ASSERT_EQ("EOF", Read());
} }
TEST_F(LogTest, Fragmentation) { TEST(LogTest, Fragmentation) {
Write("small"); Write("small");
Write(BigString("medium", 50000)); Write(BigString("medium", 50000));
Write(BigString("large", 100000)); Write(BigString("large", 100000));
@ -293,9 +281,9 @@ TEST_F(LogTest, Fragmentation) {
ASSERT_EQ("EOF", Read()); ASSERT_EQ("EOF", Read());
} }
TEST_F(LogTest, MarginalTrailer) { TEST(LogTest, MarginalTrailer) {
// Make a trailer that is exactly the same length as an empty record. // Make a trailer that is exactly the same length as an empty record.
const int n = kBlockSize - 2 * kHeaderSize; const int n = kBlockSize - 2*kHeaderSize;
Write(BigString("foo", n)); Write(BigString("foo", n));
ASSERT_EQ(kBlockSize - kHeaderSize, WrittenBytes()); ASSERT_EQ(kBlockSize - kHeaderSize, WrittenBytes());
Write(""); Write("");
@ -306,9 +294,9 @@ TEST_F(LogTest, MarginalTrailer) {
ASSERT_EQ("EOF", Read()); ASSERT_EQ("EOF", Read());
} }
TEST_F(LogTest, MarginalTrailer2) { TEST(LogTest, MarginalTrailer2) {
// Make a trailer that is exactly the same length as an empty record. // Make a trailer that is exactly the same length as an empty record.
const int n = kBlockSize - 2 * kHeaderSize; const int n = kBlockSize - 2*kHeaderSize;
Write(BigString("foo", n)); Write(BigString("foo", n));
ASSERT_EQ(kBlockSize - kHeaderSize, WrittenBytes()); ASSERT_EQ(kBlockSize - kHeaderSize, WrittenBytes());
Write("bar"); Write("bar");
@ -319,8 +307,8 @@ TEST_F(LogTest, MarginalTrailer2) {
ASSERT_EQ("", ReportMessage()); ASSERT_EQ("", ReportMessage());
} }
TEST_F(LogTest, ShortTrailer) { TEST(LogTest, ShortTrailer) {
const int n = kBlockSize - 2 * kHeaderSize + 4; const int n = kBlockSize - 2*kHeaderSize + 4;
Write(BigString("foo", n)); Write(BigString("foo", n));
ASSERT_EQ(kBlockSize - kHeaderSize + 4, WrittenBytes()); ASSERT_EQ(kBlockSize - kHeaderSize + 4, WrittenBytes());
Write(""); Write("");
@ -331,15 +319,15 @@ TEST_F(LogTest, ShortTrailer) {
ASSERT_EQ("EOF", Read()); ASSERT_EQ("EOF", Read());
} }
TEST_F(LogTest, AlignedEof) { TEST(LogTest, AlignedEof) {
const int n = kBlockSize - 2 * kHeaderSize + 4; const int n = kBlockSize - 2*kHeaderSize + 4;
Write(BigString("foo", n)); Write(BigString("foo", n));
ASSERT_EQ(kBlockSize - kHeaderSize + 4, WrittenBytes()); ASSERT_EQ(kBlockSize - kHeaderSize + 4, WrittenBytes());
ASSERT_EQ(BigString("foo", n), Read()); ASSERT_EQ(BigString("foo", n), Read());
ASSERT_EQ("EOF", Read()); ASSERT_EQ("EOF", Read());
} }
TEST_F(LogTest, OpenForAppend) { TEST(LogTest, OpenForAppend) {
Write("hello"); Write("hello");
ReopenForAppend(); ReopenForAppend();
Write("world"); Write("world");
@ -348,7 +336,7 @@ TEST_F(LogTest, OpenForAppend) {
ASSERT_EQ("EOF", Read()); ASSERT_EQ("EOF", Read());
} }
TEST_F(LogTest, RandomRead) { TEST(LogTest, RandomRead) {
const int N = 500; const int N = 500;
Random write_rnd(301); Random write_rnd(301);
for (int i = 0; i < N; i++) { for (int i = 0; i < N; i++) {
@ -363,7 +351,7 @@ TEST_F(LogTest, RandomRead) {
// Tests of all the error paths in log_reader.cc follow: // Tests of all the error paths in log_reader.cc follow:
TEST_F(LogTest, ReadError) { TEST(LogTest, ReadError) {
Write("foo"); Write("foo");
ForceError(); ForceError();
ASSERT_EQ("EOF", Read()); ASSERT_EQ("EOF", Read());
@ -371,7 +359,7 @@ TEST_F(LogTest, ReadError) {
ASSERT_EQ("OK", MatchError("read error")); ASSERT_EQ("OK", MatchError("read error"));
} }
TEST_F(LogTest, BadRecordType) { TEST(LogTest, BadRecordType) {
Write("foo"); Write("foo");
// Type is stored in header[6] // Type is stored in header[6]
IncrementByte(6, 100); IncrementByte(6, 100);
@ -381,16 +369,16 @@ TEST_F(LogTest, BadRecordType) {
ASSERT_EQ("OK", MatchError("unknown record type")); ASSERT_EQ("OK", MatchError("unknown record type"));
} }
TEST_F(LogTest, TruncatedTrailingRecordIsIgnored) { TEST(LogTest, TruncatedTrailingRecordIsIgnored) {
Write("foo"); Write("foo");
ShrinkSize(4); // Drop all payload as well as a header byte ShrinkSize(4); // Drop all payload as well as a header byte
ASSERT_EQ("EOF", Read()); ASSERT_EQ("EOF", Read());
// Truncated last record is ignored, not treated as an error. // Truncated last record is ignored, not treated as an error.
ASSERT_EQ(0, DroppedBytes()); ASSERT_EQ(0, DroppedBytes());
ASSERT_EQ("", ReportMessage()); ASSERT_EQ("", ReportMessage());
} }
TEST_F(LogTest, BadLength) { TEST(LogTest, BadLength) {
const int kPayloadSize = kBlockSize - kHeaderSize; const int kPayloadSize = kBlockSize - kHeaderSize;
Write(BigString("bar", kPayloadSize)); Write(BigString("bar", kPayloadSize));
Write("foo"); Write("foo");
@ -401,7 +389,7 @@ TEST_F(LogTest, BadLength) {
ASSERT_EQ("OK", MatchError("bad record length")); ASSERT_EQ("OK", MatchError("bad record length"));
} }
TEST_F(LogTest, BadLengthAtEndIsIgnored) { TEST(LogTest, BadLengthAtEndIsIgnored) {
Write("foo"); Write("foo");
ShrinkSize(1); ShrinkSize(1);
ASSERT_EQ("EOF", Read()); ASSERT_EQ("EOF", Read());
@ -409,7 +397,7 @@ TEST_F(LogTest, BadLengthAtEndIsIgnored) {
ASSERT_EQ("", ReportMessage()); ASSERT_EQ("", ReportMessage());
} }
TEST_F(LogTest, ChecksumMismatch) { TEST(LogTest, ChecksumMismatch) {
Write("foo"); Write("foo");
IncrementByte(0, 10); IncrementByte(0, 10);
ASSERT_EQ("EOF", Read()); ASSERT_EQ("EOF", Read());
@ -417,7 +405,7 @@ TEST_F(LogTest, ChecksumMismatch) {
ASSERT_EQ("OK", MatchError("checksum mismatch")); ASSERT_EQ("OK", MatchError("checksum mismatch"));
} }
TEST_F(LogTest, UnexpectedMiddleType) { TEST(LogTest, UnexpectedMiddleType) {
Write("foo"); Write("foo");
SetByte(6, kMiddleType); SetByte(6, kMiddleType);
FixChecksum(0, 3); FixChecksum(0, 3);
@ -426,7 +414,7 @@ TEST_F(LogTest, UnexpectedMiddleType) {
ASSERT_EQ("OK", MatchError("missing start")); ASSERT_EQ("OK", MatchError("missing start"));
} }
TEST_F(LogTest, UnexpectedLastType) { TEST(LogTest, UnexpectedLastType) {
Write("foo"); Write("foo");
SetByte(6, kLastType); SetByte(6, kLastType);
FixChecksum(0, 3); FixChecksum(0, 3);
@ -435,7 +423,7 @@ TEST_F(LogTest, UnexpectedLastType) {
ASSERT_EQ("OK", MatchError("missing start")); ASSERT_EQ("OK", MatchError("missing start"));
} }
TEST_F(LogTest, UnexpectedFullType) { TEST(LogTest, UnexpectedFullType) {
Write("foo"); Write("foo");
Write("bar"); Write("bar");
SetByte(6, kFirstType); SetByte(6, kFirstType);
@ -446,7 +434,7 @@ TEST_F(LogTest, UnexpectedFullType) {
ASSERT_EQ("OK", MatchError("partial record without end")); ASSERT_EQ("OK", MatchError("partial record without end"));
} }
TEST_F(LogTest, UnexpectedFirstType) { TEST(LogTest, UnexpectedFirstType) {
Write("foo"); Write("foo");
Write(BigString("bar", 100000)); Write(BigString("bar", 100000));
SetByte(6, kFirstType); SetByte(6, kFirstType);
@ -457,7 +445,7 @@ TEST_F(LogTest, UnexpectedFirstType) {
ASSERT_EQ("OK", MatchError("partial record without end")); ASSERT_EQ("OK", MatchError("partial record without end"));
} }
TEST_F(LogTest, MissingLastIsIgnored) { TEST(LogTest, MissingLastIsIgnored) {
Write(BigString("bar", kBlockSize)); Write(BigString("bar", kBlockSize));
// Remove the LAST block, including header. // Remove the LAST block, including header.
ShrinkSize(14); ShrinkSize(14);
@ -466,7 +454,7 @@ TEST_F(LogTest, MissingLastIsIgnored) {
ASSERT_EQ(0, DroppedBytes()); ASSERT_EQ(0, DroppedBytes());
} }
TEST_F(LogTest, PartialLastIsIgnored) { TEST(LogTest, PartialLastIsIgnored) {
Write(BigString("bar", kBlockSize)); Write(BigString("bar", kBlockSize));
// Cause a bad record length in the LAST block. // Cause a bad record length in the LAST block.
ShrinkSize(1); ShrinkSize(1);
@ -475,23 +463,7 @@ TEST_F(LogTest, PartialLastIsIgnored) {
ASSERT_EQ(0, DroppedBytes()); ASSERT_EQ(0, DroppedBytes());
} }
TEST_F(LogTest, SkipIntoMultiRecord) { TEST(LogTest, ErrorJoinsRecords) {
// Consider a fragmented record:
// first(R1), middle(R1), last(R1), first(R2)
// If initial_offset points to a record after first(R1) but before first(R2)
// incomplete fragment errors are not actual errors, and must be suppressed
// until a new first or full record is encountered.
Write(BigString("foo", 3 * kBlockSize));
Write("correct");
StartReadingAt(kBlockSize);
ASSERT_EQ("correct", Read());
ASSERT_EQ("", ReportMessage());
ASSERT_EQ(0, DroppedBytes());
ASSERT_EQ("EOF", Read());
}
TEST_F(LogTest, ErrorJoinsRecords) {
// Consider two fragmented records: // Consider two fragmented records:
// first(R1) last(R1) first(R2) last(R2) // first(R1) last(R1) first(R2) last(R2)
// where the middle two fragments disappear. We do not want // where the middle two fragments disappear. We do not want
@ -503,56 +475,74 @@ TEST_F(LogTest, ErrorJoinsRecords) {
Write("correct"); Write("correct");
// Wipe the middle block // Wipe the middle block
for (int offset = kBlockSize; offset < 2 * kBlockSize; offset++) { for (int offset = kBlockSize; offset < 2*kBlockSize; offset++) {
SetByte(offset, 'x'); SetByte(offset, 'x');
} }
ASSERT_EQ("correct", Read()); ASSERT_EQ("correct", Read());
ASSERT_EQ("EOF", Read()); ASSERT_EQ("EOF", Read());
const size_t dropped = DroppedBytes(); const size_t dropped = DroppedBytes();
ASSERT_LE(dropped, 2 * kBlockSize + 100); ASSERT_LE(dropped, 2*kBlockSize + 100);
ASSERT_GE(dropped, 2 * kBlockSize); ASSERT_GE(dropped, 2*kBlockSize);
} }
TEST_F(LogTest, ReadStart) { CheckInitialOffsetRecord(0, 0); } TEST(LogTest, ReadStart) {
CheckInitialOffsetRecord(0, 0);
}
TEST_F(LogTest, ReadSecondOneOff) { CheckInitialOffsetRecord(1, 1); } TEST(LogTest, ReadSecondOneOff) {
CheckInitialOffsetRecord(1, 1);
}
TEST_F(LogTest, ReadSecondTenThousand) { CheckInitialOffsetRecord(10000, 1); } TEST(LogTest, ReadSecondTenThousand) {
CheckInitialOffsetRecord(10000, 1);
}
TEST_F(LogTest, ReadSecondStart) { CheckInitialOffsetRecord(10007, 1); } TEST(LogTest, ReadSecondStart) {
CheckInitialOffsetRecord(10007, 1);
}
TEST_F(LogTest, ReadThirdOneOff) { CheckInitialOffsetRecord(10008, 2); } TEST(LogTest, ReadThirdOneOff) {
CheckInitialOffsetRecord(10008, 2);
}
TEST_F(LogTest, ReadThirdStart) { CheckInitialOffsetRecord(20014, 2); } TEST(LogTest, ReadThirdStart) {
CheckInitialOffsetRecord(20014, 2);
}
TEST_F(LogTest, ReadFourthOneOff) { CheckInitialOffsetRecord(20015, 3); } TEST(LogTest, ReadFourthOneOff) {
CheckInitialOffsetRecord(20015, 3);
}
TEST_F(LogTest, ReadFourthFirstBlockTrailer) { TEST(LogTest, ReadFourthFirstBlockTrailer) {
CheckInitialOffsetRecord(log::kBlockSize - 4, 3); CheckInitialOffsetRecord(log::kBlockSize - 4, 3);
} }
TEST_F(LogTest, ReadFourthMiddleBlock) { TEST(LogTest, ReadFourthMiddleBlock) {
CheckInitialOffsetRecord(log::kBlockSize + 1, 3); CheckInitialOffsetRecord(log::kBlockSize + 1, 3);
} }
TEST_F(LogTest, ReadFourthLastBlock) { TEST(LogTest, ReadFourthLastBlock) {
CheckInitialOffsetRecord(2 * log::kBlockSize + 1, 3); CheckInitialOffsetRecord(2 * log::kBlockSize + 1, 3);
} }
TEST_F(LogTest, ReadFourthStart) { TEST(LogTest, ReadFourthStart) {
CheckInitialOffsetRecord( CheckInitialOffsetRecord(
2 * (kHeaderSize + 1000) + (2 * log::kBlockSize - 1000) + 3 * kHeaderSize, 2 * (kHeaderSize + 1000) + (2 * log::kBlockSize - 1000) + 3 * kHeaderSize,
3); 3);
} }
TEST_F(LogTest, ReadInitialOffsetIntoBlockPadding) { TEST(LogTest, ReadEnd) {
CheckInitialOffsetRecord(3 * log::kBlockSize - 3, 5); CheckOffsetPastEndReturnsNoRecords(0);
} }
TEST_F(LogTest, ReadEnd) { CheckOffsetPastEndReturnsNoRecords(0); } TEST(LogTest, ReadPastEnd) {
CheckOffsetPastEndReturnsNoRecords(5);
TEST_F(LogTest, ReadPastEnd) { CheckOffsetPastEndReturnsNoRecords(5); } }
} // namespace log } // namespace log
} // namespace leveldb } // namespace leveldb
int main(int argc, char** argv) {
return leveldb::test::RunAllTests();
}

View File

@ -4,8 +4,7 @@
#include "db/log_writer.h" #include "db/log_writer.h"
#include <cstdint> #include <stdint.h>
#include "leveldb/env.h" #include "leveldb/env.h"
#include "util/coding.h" #include "util/coding.h"
#include "util/crc32c.h" #include "util/crc32c.h"
@ -20,7 +19,9 @@ static void InitTypeCrc(uint32_t* type_crc) {
} }
} }
Writer::Writer(WritableFile* dest) : dest_(dest), block_offset_(0) { Writer::Writer(WritableFile* dest)
: dest_(dest),
block_offset_(0) {
InitTypeCrc(type_crc_); InitTypeCrc(type_crc_);
} }
@ -29,7 +30,8 @@ Writer::Writer(WritableFile* dest, uint64_t dest_length)
InitTypeCrc(type_crc_); InitTypeCrc(type_crc_);
} }
Writer::~Writer() = default; Writer::~Writer() {
}
Status Writer::AddRecord(const Slice& slice) { Status Writer::AddRecord(const Slice& slice) {
const char* ptr = slice.data(); const char* ptr = slice.data();
@ -47,7 +49,7 @@ Status Writer::AddRecord(const Slice& slice) {
// Switch to a new block // Switch to a new block
if (leftover > 0) { if (leftover > 0) {
// Fill the trailer (literal below relies on kHeaderSize being 7) // Fill the trailer (literal below relies on kHeaderSize being 7)
static_assert(kHeaderSize == 7, ""); assert(kHeaderSize == 7);
dest_->Append(Slice("\x00\x00\x00\x00\x00\x00", leftover)); dest_->Append(Slice("\x00\x00\x00\x00\x00\x00", leftover));
} }
block_offset_ = 0; block_offset_ = 0;
@ -79,31 +81,30 @@ Status Writer::AddRecord(const Slice& slice) {
return s; return s;
} }
Status Writer::EmitPhysicalRecord(RecordType t, const char* ptr, Status Writer::EmitPhysicalRecord(RecordType t, const char* ptr, size_t n) {
size_t length) { assert(n <= 0xffff); // Must fit in two bytes
assert(length <= 0xffff); // Must fit in two bytes assert(block_offset_ + kHeaderSize + n <= kBlockSize);
assert(block_offset_ + kHeaderSize + length <= kBlockSize);
// Format the header // Format the header
char buf[kHeaderSize]; char buf[kHeaderSize];
buf[4] = static_cast<char>(length & 0xff); buf[4] = static_cast<char>(n & 0xff);
buf[5] = static_cast<char>(length >> 8); buf[5] = static_cast<char>(n >> 8);
buf[6] = static_cast<char>(t); buf[6] = static_cast<char>(t);
// Compute the crc of the record type and the payload. // Compute the crc of the record type and the payload.
uint32_t crc = crc32c::Extend(type_crc_[t], ptr, length); uint32_t crc = crc32c::Extend(type_crc_[t], ptr, n);
crc = crc32c::Mask(crc); // Adjust for storage crc = crc32c::Mask(crc); // Adjust for storage
EncodeFixed32(buf, crc); EncodeFixed32(buf, crc);
// Write the header and the payload // Write the header and the payload
Status s = dest_->Append(Slice(buf, kHeaderSize)); Status s = dest_->Append(Slice(buf, kHeaderSize));
if (s.ok()) { if (s.ok()) {
s = dest_->Append(Slice(ptr, length)); s = dest_->Append(Slice(ptr, n));
if (s.ok()) { if (s.ok()) {
s = dest_->Flush(); s = dest_->Flush();
} }
} }
block_offset_ += kHeaderSize + length; block_offset_ += kHeaderSize + n;
return s; return s;
} }

View File

@ -5,8 +5,7 @@
#ifndef STORAGE_LEVELDB_DB_LOG_WRITER_H_ #ifndef STORAGE_LEVELDB_DB_LOG_WRITER_H_
#define STORAGE_LEVELDB_DB_LOG_WRITER_H_ #define STORAGE_LEVELDB_DB_LOG_WRITER_H_
#include <cstdint> #include <stdint.h>
#include "db/log_format.h" #include "db/log_format.h"
#include "leveldb/slice.h" #include "leveldb/slice.h"
#include "leveldb/status.h" #include "leveldb/status.h"
@ -29,23 +28,24 @@ class Writer {
// "*dest" must remain live while this Writer is in use. // "*dest" must remain live while this Writer is in use.
Writer(WritableFile* dest, uint64_t dest_length); Writer(WritableFile* dest, uint64_t dest_length);
Writer(const Writer&) = delete;
Writer& operator=(const Writer&) = delete;
~Writer(); ~Writer();
Status AddRecord(const Slice& slice); Status AddRecord(const Slice& slice);
private: private:
Status EmitPhysicalRecord(RecordType type, const char* ptr, size_t length);
WritableFile* dest_; WritableFile* dest_;
int block_offset_; // Current offset in block int block_offset_; // Current offset in block
// crc32c values for all supported record types. These are // crc32c values for all supported record types. These are
// pre-computed to reduce the overhead of computing the crc of the // pre-computed to reduce the overhead of computing the crc of the
// record type stored in the header. // record type stored in the header.
uint32_t type_crc_[kMaxRecordType + 1]; uint32_t type_crc_[kMaxRecordType + 1];
Status EmitPhysicalRecord(RecordType type, const char* ptr, size_t length);
// No copying allowed
Writer(const Writer&);
void operator=(const Writer&);
}; };
} // namespace log } // namespace log

View File

@ -18,15 +18,20 @@ static Slice GetLengthPrefixedSlice(const char* data) {
return Slice(p, len); return Slice(p, len);
} }
MemTable::MemTable(const InternalKeyComparator& comparator) MemTable::MemTable(const InternalKeyComparator& cmp)
: comparator_(comparator), refs_(0), table_(comparator_, &arena_) {} : comparator_(cmp),
refs_(0),
table_(comparator_, &arena_) {
}
MemTable::~MemTable() { assert(refs_ == 0); } MemTable::~MemTable() {
assert(refs_ == 0);
}
size_t MemTable::ApproximateMemoryUsage() { return arena_.MemoryUsage(); } size_t MemTable::ApproximateMemoryUsage() { return arena_.MemoryUsage(); }
int MemTable::KeyComparator::operator()(const char* aptr, int MemTable::KeyComparator::operator()(const char* aptr, const char* bptr)
const char* bptr) const { const {
// Internal keys are encoded as length-prefixed strings. // Internal keys are encoded as length-prefixed strings.
Slice a = GetLengthPrefixedSlice(aptr); Slice a = GetLengthPrefixedSlice(aptr);
Slice b = GetLengthPrefixedSlice(bptr); Slice b = GetLengthPrefixedSlice(bptr);
@ -43,59 +48,60 @@ static const char* EncodeKey(std::string* scratch, const Slice& target) {
return scratch->data(); return scratch->data();
} }
class MemTableIterator : public Iterator { class MemTableIterator: public Iterator {
public: public:
explicit MemTableIterator(MemTable::Table* table) : iter_(table) {} explicit MemTableIterator(MemTable::Table* table) : iter_(table) { }
MemTableIterator(const MemTableIterator&) = delete; virtual bool Valid() const { return iter_.Valid(); }
MemTableIterator& operator=(const MemTableIterator&) = delete; virtual void Seek(const Slice& k) { iter_.Seek(EncodeKey(&tmp_, k)); }
virtual void SeekToFirst() { iter_.SeekToFirst(); }
~MemTableIterator() override = default; virtual void SeekToLast() { iter_.SeekToLast(); }
virtual void Next() { iter_.Next(); }
bool Valid() const override { return iter_.Valid(); } virtual void Prev() { iter_.Prev(); }
void Seek(const Slice& k) override { iter_.Seek(EncodeKey(&tmp_, k)); } virtual Slice key() const { return GetLengthPrefixedSlice(iter_.key()); }
void SeekToFirst() override { iter_.SeekToFirst(); } virtual Slice value() const {
void SeekToLast() override { iter_.SeekToLast(); }
void Next() override { iter_.Next(); }
void Prev() override { iter_.Prev(); }
Slice key() const override { return GetLengthPrefixedSlice(iter_.key()); }
Slice value() const override {
Slice key_slice = GetLengthPrefixedSlice(iter_.key()); Slice key_slice = GetLengthPrefixedSlice(iter_.key());
return GetLengthPrefixedSlice(key_slice.data() + key_slice.size()); return GetLengthPrefixedSlice(key_slice.data() + key_slice.size());
} }
Status status() const override { return Status::OK(); } virtual Status status() const { return Status::OK(); }
private: private:
MemTable::Table::Iterator iter_; MemTable::Table::Iterator iter_;
std::string tmp_; // For passing to EncodeKey std::string tmp_; // For passing to EncodeKey
// No copying allowed
MemTableIterator(const MemTableIterator&);
void operator=(const MemTableIterator&);
}; };
Iterator* MemTable::NewIterator() { return new MemTableIterator(&table_); } Iterator* MemTable::NewIterator() {
return new MemTableIterator(&table_);
}
void MemTable::Add(SequenceNumber s, ValueType type, const Slice& key, void MemTable::Add(SequenceNumber s, ValueType type,
const Slice& key,
const Slice& value) { const Slice& value) {
// Format of an entry is concatenation of: // Format of an entry is concatenation of:
// key_size : varint32 of internal_key.size() // key_size : varint32 of internal_key.size()
// key bytes : char[internal_key.size()] // key bytes : char[internal_key.size()]
// tag : uint64((sequence << 8) | type)
// value_size : varint32 of value.size() // value_size : varint32 of value.size()
// value bytes : char[value.size()] // value bytes : char[value.size()]
size_t key_size = key.size(); size_t key_size = key.size();
size_t val_size = value.size(); size_t val_size = value.size();
size_t internal_key_size = key_size + 8; size_t internal_key_size = key_size + 8;
const size_t encoded_len = VarintLength(internal_key_size) + const size_t encoded_len =
internal_key_size + VarintLength(val_size) + VarintLength(internal_key_size) + internal_key_size +
val_size; VarintLength(val_size) + val_size;
char* buf = arena_.Allocate(encoded_len); char* buf = arena_.Allocate(encoded_len);
char* p = EncodeVarint32(buf, internal_key_size); char* p = EncodeVarint32(buf, internal_key_size);
std::memcpy(p, key.data(), key_size); memcpy(p, key.data(), key_size);
p += key_size; p += key_size;
EncodeFixed64(p, (s << 8) | type); EncodeFixed64(p, (s << 8) | type);
p += 8; p += 8;
p = EncodeVarint32(p, val_size); p = EncodeVarint32(p, val_size);
std::memcpy(p, value.data(), val_size); memcpy(p, value.data(), val_size);
assert(p + val_size == buf + encoded_len); assert((p + val_size) - buf == encoded_len);
table_.Insert(buf); table_.Insert(buf);
} }
@ -115,9 +121,10 @@ bool MemTable::Get(const LookupKey& key, std::string* value, Status* s) {
// all entries with overly large sequence numbers. // all entries with overly large sequence numbers.
const char* entry = iter.key(); const char* entry = iter.key();
uint32_t key_length; uint32_t key_length;
const char* key_ptr = GetVarint32Ptr(entry, entry + 5, &key_length); const char* key_ptr = GetVarint32Ptr(entry, entry+5, &key_length);
if (comparator_.comparator.user_comparator()->Compare( if (comparator_.comparator.user_comparator()->Compare(
Slice(key_ptr, key_length - 8), key.user_key()) == 0) { Slice(key_ptr, key_length - 8),
key.user_key()) == 0) {
// Correct user key // Correct user key
const uint64_t tag = DecodeFixed64(key_ptr + key_length - 8); const uint64_t tag = DecodeFixed64(key_ptr + key_length - 8);
switch (static_cast<ValueType>(tag & 0xff)) { switch (static_cast<ValueType>(tag & 0xff)) {

View File

@ -6,15 +6,15 @@
#define STORAGE_LEVELDB_DB_MEMTABLE_H_ #define STORAGE_LEVELDB_DB_MEMTABLE_H_
#include <string> #include <string>
#include "leveldb/db.h"
#include "db/dbformat.h" #include "db/dbformat.h"
#include "db/skiplist.h" #include "db/skiplist.h"
#include "leveldb/db.h"
#include "util/arena.h" #include "util/arena.h"
namespace leveldb { namespace leveldb {
class InternalKeyComparator; class InternalKeyComparator;
class Mutex;
class MemTableIterator; class MemTableIterator;
class MemTable { class MemTable {
@ -23,9 +23,6 @@ class MemTable {
// is zero and the caller must call Ref() at least once. // is zero and the caller must call Ref() at least once.
explicit MemTable(const InternalKeyComparator& comparator); explicit MemTable(const InternalKeyComparator& comparator);
MemTable(const MemTable&) = delete;
MemTable& operator=(const MemTable&) = delete;
// Increase reference count. // Increase reference count.
void Ref() { ++refs_; } void Ref() { ++refs_; }
@ -39,7 +36,10 @@ class MemTable {
} }
// Returns an estimate of the number of bytes of data in use by this // Returns an estimate of the number of bytes of data in use by this
// data structure. It is safe to call when MemTable is being modified. // data structure.
//
// REQUIRES: external synchronization to prevent simultaneous
// operations on the same MemTable.
size_t ApproximateMemoryUsage(); size_t ApproximateMemoryUsage();
// Return an iterator that yields the contents of the memtable. // Return an iterator that yields the contents of the memtable.
@ -53,7 +53,8 @@ class MemTable {
// Add an entry into memtable that maps key to value at the // Add an entry into memtable that maps key to value at the
// specified sequence number and with the specified type. // specified sequence number and with the specified type.
// Typically value will be empty if type==kTypeDeletion. // Typically value will be empty if type==kTypeDeletion.
void Add(SequenceNumber seq, ValueType type, const Slice& key, void Add(SequenceNumber seq, ValueType type,
const Slice& key,
const Slice& value); const Slice& value);
// If memtable contains a value for key, store it in *value and return true. // If memtable contains a value for key, store it in *value and return true.
@ -63,23 +64,26 @@ class MemTable {
bool Get(const LookupKey& key, std::string* value, Status* s); bool Get(const LookupKey& key, std::string* value, Status* s);
private: private:
friend class MemTableIterator; ~MemTable(); // Private since only Unref() should be used to delete it
friend class MemTableBackwardIterator;
struct KeyComparator { struct KeyComparator {
const InternalKeyComparator comparator; const InternalKeyComparator comparator;
explicit KeyComparator(const InternalKeyComparator& c) : comparator(c) {} explicit KeyComparator(const InternalKeyComparator& c) : comparator(c) { }
int operator()(const char* a, const char* b) const; int operator()(const char* a, const char* b) const;
}; };
friend class MemTableIterator;
friend class MemTableBackwardIterator;
typedef SkipList<const char*, KeyComparator> Table; typedef SkipList<const char*, KeyComparator> Table;
~MemTable(); // Private since only Unref() should be used to delete it
KeyComparator comparator_; KeyComparator comparator_;
int refs_; int refs_;
Arena arena_; Arena arena_;
Table table_; Table table_;
// No copying allowed
MemTable(const MemTable&);
void operator=(const MemTable&);
}; };
} // namespace leveldb } // namespace leveldb

View File

@ -2,7 +2,6 @@
// Use of this source code is governed by a BSD-style license that can be // Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file. See the AUTHORS file for names of contributors. // found in the LICENSE file. See the AUTHORS file for names of contributors.
#include "gtest/gtest.h"
#include "db/db_impl.h" #include "db/db_impl.h"
#include "db/filename.h" #include "db/filename.h"
#include "db/version_set.h" #include "db/version_set.h"
@ -11,14 +10,15 @@
#include "leveldb/env.h" #include "leveldb/env.h"
#include "leveldb/write_batch.h" #include "leveldb/write_batch.h"
#include "util/logging.h" #include "util/logging.h"
#include "util/testharness.h"
#include "util/testutil.h" #include "util/testutil.h"
namespace leveldb { namespace leveldb {
class RecoveryTest : public testing::Test { class RecoveryTest {
public: public:
RecoveryTest() : env_(Env::Default()), db_(nullptr) { RecoveryTest() : env_(Env::Default()), db_(NULL) {
dbname_ = testing::TempDir() + "recovery_test"; dbname_ = test::TmpDir() + "/recovery_test";
DestroyDB(dbname_, Options()); DestroyDB(dbname_, Options());
Open(); Open();
} }
@ -44,26 +44,22 @@ class RecoveryTest : public testing::Test {
void Close() { void Close() {
delete db_; delete db_;
db_ = nullptr; db_ = NULL;
} }
Status OpenWithStatus(Options* options = nullptr) { void Open(Options* options = NULL) {
Close(); Close();
Options opts; Options opts;
if (options != nullptr) { if (options != NULL) {
opts = *options; opts = *options;
} else { } else {
opts.reuse_logs = true; // TODO(sanjay): test both ways opts.reuse_logs = true; // TODO(sanjay): test both ways
opts.create_if_missing = true; opts.create_if_missing = true;
} }
if (opts.env == nullptr) { if (opts.env == NULL) {
opts.env = env_; opts.env = env_;
} }
return DB::Open(opts, dbname_, &db_); ASSERT_OK(DB::Open(opts, dbname_, &db_));
}
void Open(Options* options = nullptr) {
ASSERT_LEVELDB_OK(OpenWithStatus(options));
ASSERT_EQ(1, NumLogs()); ASSERT_EQ(1, NumLogs());
} }
@ -71,7 +67,7 @@ class RecoveryTest : public testing::Test {
return db_->Put(WriteOptions(), k, v); return db_->Put(WriteOptions(), k, v);
} }
std::string Get(const std::string& k, const Snapshot* snapshot = nullptr) { std::string Get(const std::string& k, const Snapshot* snapshot = NULL) {
std::string result; std::string result;
Status s = db_->Get(ReadOptions(), k, &result); Status s = db_->Get(ReadOptions(), k, &result);
if (s.IsNotFound()) { if (s.IsNotFound()) {
@ -84,38 +80,34 @@ class RecoveryTest : public testing::Test {
std::string ManifestFileName() { std::string ManifestFileName() {
std::string current; std::string current;
EXPECT_LEVELDB_OK( ASSERT_OK(ReadFileToString(env_, CurrentFileName(dbname_), &current));
ReadFileToString(env_, CurrentFileName(dbname_), &current));
size_t len = current.size(); size_t len = current.size();
if (len > 0 && current[len - 1] == '\n') { if (len > 0 && current[len-1] == '\n') {
current.resize(len - 1); current.resize(len - 1);
} }
return dbname_ + "/" + current; return dbname_ + "/" + current;
} }
std::string LogName(uint64_t number) { return LogFileName(dbname_, number); } std::string LogName(uint64_t number) {
return LogFileName(dbname_, number);
}
size_t RemoveLogFiles() { size_t DeleteLogFiles() {
// Linux allows unlinking open files, but Windows does not.
// Closing the db allows for file deletion.
Close();
std::vector<uint64_t> logs = GetFiles(kLogFile); std::vector<uint64_t> logs = GetFiles(kLogFile);
for (size_t i = 0; i < logs.size(); i++) { for (size_t i = 0; i < logs.size(); i++) {
EXPECT_LEVELDB_OK(env_->RemoveFile(LogName(logs[i]))) << LogName(logs[i]); ASSERT_OK(env_->DeleteFile(LogName(logs[i]))) << LogName(logs[i]);
} }
return logs.size(); return logs.size();
} }
void RemoveManifestFile() { uint64_t FirstLogFile() {
ASSERT_LEVELDB_OK(env_->RemoveFile(ManifestFileName())); return GetFiles(kLogFile)[0];
} }
uint64_t FirstLogFile() { return GetFiles(kLogFile)[0]; } std::vector<std::uint64_t> GetFiles(FileType t) {
std::vector<uint64_t> GetFiles(FileType t) {
std::vector<std::string> filenames; std::vector<std::string> filenames;
EXPECT_LEVELDB_OK(env_->GetChildren(dbname_, &filenames)); ASSERT_OK(env_->GetChildren(dbname_, &filenames));
std::vector<uint64_t> result; std::vector<std::uint64_t> result;
for (size_t i = 0; i < filenames.size(); i++) { for (size_t i = 0; i < filenames.size(); i++) {
uint64_t number; uint64_t number;
FileType type; FileType type;
@ -126,29 +118,35 @@ class RecoveryTest : public testing::Test {
return result; return result;
} }
int NumLogs() { return GetFiles(kLogFile).size(); } int NumLogs() {
return GetFiles(kLogFile).size();
}
int NumTables() { return GetFiles(kTableFile).size(); } int NumTables() {
return GetFiles(kTableFile).size();
}
uint64_t FileSize(const std::string& fname) { uint64_t FileSize(const std::string& fname) {
uint64_t result; uint64_t result;
EXPECT_LEVELDB_OK(env_->GetFileSize(fname, &result)) << fname; ASSERT_OK(env_->GetFileSize(fname, &result)) << fname;
return result; return result;
} }
void CompactMemTable() { dbfull()->TEST_CompactMemTable(); } void CompactMemTable() {
dbfull()->TEST_CompactMemTable();
}
// Directly construct a log file that sets key to val. // Directly construct a log file that sets key to val.
void MakeLogFile(uint64_t lognum, SequenceNumber seq, Slice key, Slice val) { void MakeLogFile(uint64_t lognum, SequenceNumber seq, Slice key, Slice val) {
std::string fname = LogFileName(dbname_, lognum); std::string fname = LogFileName(dbname_, lognum);
WritableFile* file; WritableFile* file;
ASSERT_LEVELDB_OK(env_->NewWritableFile(fname, &file)); ASSERT_OK(env_->NewWritableFile(fname, &file));
log::Writer writer(file); log::Writer writer(file);
WriteBatch batch; WriteBatch batch;
batch.Put(key, val); batch.Put(key, val);
WriteBatchInternal::SetSequence(&batch, seq); WriteBatchInternal::SetSequence(&batch, seq);
ASSERT_LEVELDB_OK(writer.AddRecord(WriteBatchInternal::Contents(&batch))); ASSERT_OK(writer.AddRecord(WriteBatchInternal::Contents(&batch)));
ASSERT_LEVELDB_OK(file->Flush()); ASSERT_OK(file->Flush());
delete file; delete file;
} }
@ -158,13 +156,12 @@ class RecoveryTest : public testing::Test {
DB* db_; DB* db_;
}; };
TEST_F(RecoveryTest, ManifestReused) { TEST(RecoveryTest, ManifestReused) {
if (!CanAppend()) { if (!CanAppend()) {
std::fprintf(stderr, fprintf(stderr, "skipping test because env does not support appending\n");
"skipping test because env does not support appending\n");
return; return;
} }
ASSERT_LEVELDB_OK(Put("foo", "bar")); ASSERT_OK(Put("foo", "bar"));
Close(); Close();
std::string old_manifest = ManifestFileName(); std::string old_manifest = ManifestFileName();
Open(); Open();
@ -175,13 +172,12 @@ TEST_F(RecoveryTest, ManifestReused) {
ASSERT_EQ("bar", Get("foo")); ASSERT_EQ("bar", Get("foo"));
} }
TEST_F(RecoveryTest, LargeManifestCompacted) { TEST(RecoveryTest, LargeManifestCompacted) {
if (!CanAppend()) { if (!CanAppend()) {
std::fprintf(stderr, fprintf(stderr, "skipping test because env does not support appending\n");
"skipping test because env does not support appending\n");
return; return;
} }
ASSERT_LEVELDB_OK(Put("foo", "bar")); ASSERT_OK(Put("foo", "bar"));
Close(); Close();
std::string old_manifest = ManifestFileName(); std::string old_manifest = ManifestFileName();
@ -189,10 +185,10 @@ TEST_F(RecoveryTest, LargeManifestCompacted) {
{ {
uint64_t len = FileSize(old_manifest); uint64_t len = FileSize(old_manifest);
WritableFile* file; WritableFile* file;
ASSERT_LEVELDB_OK(env()->NewAppendableFile(old_manifest, &file)); ASSERT_OK(env()->NewAppendableFile(old_manifest, &file));
std::string zeroes(3 * 1048576 - static_cast<size_t>(len), 0); std::string zeroes(3*1048576 - static_cast<size_t>(len), 0);
ASSERT_LEVELDB_OK(file->Append(zeroes)); ASSERT_OK(file->Append(zeroes));
ASSERT_LEVELDB_OK(file->Flush()); ASSERT_OK(file->Flush());
delete file; delete file;
} }
@ -207,23 +203,22 @@ TEST_F(RecoveryTest, LargeManifestCompacted) {
ASSERT_EQ("bar", Get("foo")); ASSERT_EQ("bar", Get("foo"));
} }
TEST_F(RecoveryTest, NoLogFiles) { TEST(RecoveryTest, NoLogFiles) {
ASSERT_LEVELDB_OK(Put("foo", "bar")); ASSERT_OK(Put("foo", "bar"));
ASSERT_EQ(1, RemoveLogFiles()); ASSERT_EQ(1, DeleteLogFiles());
Open(); Open();
ASSERT_EQ("NOT_FOUND", Get("foo")); ASSERT_EQ("NOT_FOUND", Get("foo"));
Open(); Open();
ASSERT_EQ("NOT_FOUND", Get("foo")); ASSERT_EQ("NOT_FOUND", Get("foo"));
} }
TEST_F(RecoveryTest, LogFileReuse) { TEST(RecoveryTest, LogFileReuse) {
if (!CanAppend()) { if (!CanAppend()) {
std::fprintf(stderr, fprintf(stderr, "skipping test because env does not support appending\n");
"skipping test because env does not support appending\n");
return; return;
} }
for (int i = 0; i < 2; i++) { for (int i = 0; i < 2; i++) {
ASSERT_LEVELDB_OK(Put("foo", "bar")); ASSERT_OK(Put("foo", "bar"));
if (i == 0) { if (i == 0) {
// Compact to ensure current log is empty // Compact to ensure current log is empty
CompactMemTable(); CompactMemTable();
@ -247,13 +242,13 @@ TEST_F(RecoveryTest, LogFileReuse) {
} }
} }
TEST_F(RecoveryTest, MultipleMemTables) { TEST(RecoveryTest, MultipleMemTables) {
// Make a large log. // Make a large log.
const int kNum = 1000; const int kNum = 1000;
for (int i = 0; i < kNum; i++) { for (int i = 0; i < kNum; i++) {
char buf[100]; char buf[100];
std::snprintf(buf, sizeof(buf), "%050d", i); snprintf(buf, sizeof(buf), "%050d", i);
ASSERT_LEVELDB_OK(Put(buf, buf)); ASSERT_OK(Put(buf, buf));
} }
ASSERT_EQ(0, NumTables()); ASSERT_EQ(0, NumTables());
Close(); Close();
@ -264,35 +259,35 @@ TEST_F(RecoveryTest, MultipleMemTables) {
// Force creation of multiple memtables by reducing the write buffer size. // Force creation of multiple memtables by reducing the write buffer size.
Options opt; Options opt;
opt.reuse_logs = true; opt.reuse_logs = true;
opt.write_buffer_size = (kNum * 100) / 2; opt.write_buffer_size = (kNum*100) / 2;
Open(&opt); Open(&opt);
ASSERT_LE(2, NumTables()); ASSERT_LE(2, NumTables());
ASSERT_EQ(1, NumLogs()); ASSERT_EQ(1, NumLogs());
ASSERT_NE(old_log_file, FirstLogFile()) << "must not reuse log"; ASSERT_NE(old_log_file, FirstLogFile()) << "must not reuse log";
for (int i = 0; i < kNum; i++) { for (int i = 0; i < kNum; i++) {
char buf[100]; char buf[100];
std::snprintf(buf, sizeof(buf), "%050d", i); snprintf(buf, sizeof(buf), "%050d", i);
ASSERT_EQ(buf, Get(buf)); ASSERT_EQ(buf, Get(buf));
} }
} }
TEST_F(RecoveryTest, MultipleLogFiles) { TEST(RecoveryTest, MultipleLogFiles) {
ASSERT_LEVELDB_OK(Put("foo", "bar")); ASSERT_OK(Put("foo", "bar"));
Close(); Close();
ASSERT_EQ(1, NumLogs()); ASSERT_EQ(1, NumLogs());
// Make a bunch of uncompacted log files. // Make a bunch of uncompacted log files.
uint64_t old_log = FirstLogFile(); uint64_t old_log = FirstLogFile();
MakeLogFile(old_log + 1, 1000, "hello", "world"); MakeLogFile(old_log+1, 1000, "hello", "world");
MakeLogFile(old_log + 2, 1001, "hi", "there"); MakeLogFile(old_log+2, 1001, "hi", "there");
MakeLogFile(old_log + 3, 1002, "foo", "bar2"); MakeLogFile(old_log+3, 1002, "foo", "bar2");
// Recover and check that all log files were processed. // Recover and check that all log files were processed.
Open(); Open();
ASSERT_LE(1, NumTables()); ASSERT_LE(1, NumTables());
ASSERT_EQ(1, NumLogs()); ASSERT_EQ(1, NumLogs());
uint64_t new_log = FirstLogFile(); uint64_t new_log = FirstLogFile();
ASSERT_LE(old_log + 3, new_log); ASSERT_LE(old_log+3, new_log);
ASSERT_EQ("bar2", Get("foo")); ASSERT_EQ("bar2", Get("foo"));
ASSERT_EQ("world", Get("hello")); ASSERT_EQ("world", Get("hello"));
ASSERT_EQ("there", Get("hi")); ASSERT_EQ("there", Get("hi"));
@ -310,7 +305,7 @@ TEST_F(RecoveryTest, MultipleLogFiles) {
// Check that introducing an older log file does not cause it to be re-read. // Check that introducing an older log file does not cause it to be re-read.
Close(); Close();
MakeLogFile(old_log + 1, 2000, "hello", "stale write"); MakeLogFile(old_log+1, 2000, "hello", "stale write");
Open(); Open();
ASSERT_LE(1, NumTables()); ASSERT_LE(1, NumTables());
ASSERT_EQ(1, NumLogs()); ASSERT_EQ(1, NumLogs());
@ -322,18 +317,8 @@ TEST_F(RecoveryTest, MultipleLogFiles) {
ASSERT_EQ("there", Get("hi")); ASSERT_EQ("there", Get("hi"));
} }
TEST_F(RecoveryTest, ManifestMissing) {
ASSERT_LEVELDB_OK(Put("foo", "bar"));
Close();
RemoveManifestFile();
Status status = OpenWithStatus();
#if defined(LEVELDB_PLATFORM_CHROMIUM)
// TODO(crbug.com/760362): See comment in MakeIOError() from env_chromium.cc.
ASSERT_TRUE(status.IsIOError());
#else
ASSERT_TRUE(status.IsCorruption());
#endif // defined(LEVELDB_PLATFORM_CHROMIUM)
}
} // namespace leveldb } // namespace leveldb
int main(int argc, char** argv) {
return leveldb::test::RunAllTests();
}

View File

@ -54,7 +54,7 @@ class Repairer {
owns_cache_(options_.block_cache != options.block_cache), owns_cache_(options_.block_cache != options.block_cache),
next_file_number_(1) { next_file_number_(1) {
// TableCache can be small since we expect each table to be opened once. // TableCache can be small since we expect each table to be opened once.
table_cache_ = new TableCache(dbname_, options_, 10); table_cache_ = new TableCache(dbname_, &options_, 10);
} }
~Repairer() { ~Repairer() {
@ -84,7 +84,9 @@ class Repairer {
"recovered %d files; %llu bytes. " "recovered %d files; %llu bytes. "
"Some data may have been lost. " "Some data may have been lost. "
"****", "****",
dbname_.c_str(), static_cast<int>(tables_.size()), bytes); dbname_.c_str(),
static_cast<int>(tables_.size()),
bytes);
} }
return status; return status;
} }
@ -95,6 +97,22 @@ class Repairer {
SequenceNumber max_sequence; SequenceNumber max_sequence;
}; };
std::string const dbname_;
Env* const env_;
InternalKeyComparator const icmp_;
InternalFilterPolicy const ipolicy_;
Options const options_;
bool owns_info_log_;
bool owns_cache_;
TableCache* table_cache_;
VersionEdit edit_;
std::vector<std::string> manifests_;
std::vector<uint64_t> table_numbers_;
std::vector<uint64_t> logs_;
std::vector<TableInfo> tables_;
uint64_t next_file_number_;
Status FindFiles() { Status FindFiles() {
std::vector<std::string> filenames; std::vector<std::string> filenames;
Status status = env_->GetChildren(dbname_, &filenames); Status status = env_->GetChildren(dbname_, &filenames);
@ -134,7 +152,8 @@ class Repairer {
Status status = ConvertLogToTable(logs_[i]); Status status = ConvertLogToTable(logs_[i]);
if (!status.ok()) { if (!status.ok()) {
Log(options_.info_log, "Log #%llu: ignoring conversion error: %s", Log(options_.info_log, "Log #%llu: ignoring conversion error: %s",
(unsigned long long)logs_[i], status.ToString().c_str()); (unsigned long long) logs_[i],
status.ToString().c_str());
} }
ArchiveFile(logname); ArchiveFile(logname);
} }
@ -145,10 +164,11 @@ class Repairer {
Env* env; Env* env;
Logger* info_log; Logger* info_log;
uint64_t lognum; uint64_t lognum;
void Corruption(size_t bytes, const Status& s) override { virtual void Corruption(size_t bytes, const Status& s) {
// We print error messages for corruption, but continue repairing. // We print error messages for corruption, but continue repairing.
Log(info_log, "Log #%llu: dropping %d bytes; %s", Log(info_log, "Log #%llu: dropping %d bytes; %s",
(unsigned long long)lognum, static_cast<int>(bytes), (unsigned long long) lognum,
static_cast<int>(bytes),
s.ToString().c_str()); s.ToString().c_str());
} }
}; };
@ -170,8 +190,8 @@ class Repairer {
// corruptions cause entire commits to be skipped instead of // corruptions cause entire commits to be skipped instead of
// propagating bad information (like overly large sequence // propagating bad information (like overly large sequence
// numbers). // numbers).
log::Reader reader(lfile, &reporter, false /*do not checksum*/, log::Reader reader(lfile, &reporter, false/*do not checksum*/,
0 /*initial_offset*/); 0/*initial_offset*/);
// Read all the records and add to a memtable // Read all the records and add to a memtable
std::string scratch; std::string scratch;
@ -182,8 +202,8 @@ class Repairer {
int counter = 0; int counter = 0;
while (reader.ReadRecord(&record, &scratch)) { while (reader.ReadRecord(&record, &scratch)) {
if (record.size() < 12) { if (record.size() < 12) {
reporter.Corruption(record.size(), reporter.Corruption(
Status::Corruption("log record too small")); record.size(), Status::Corruption("log record too small"));
continue; continue;
} }
WriteBatchInternal::SetContents(&batch, record); WriteBatchInternal::SetContents(&batch, record);
@ -192,7 +212,8 @@ class Repairer {
counter += WriteBatchInternal::Count(&batch); counter += WriteBatchInternal::Count(&batch);
} else { } else {
Log(options_.info_log, "Log #%llu: ignoring %s", Log(options_.info_log, "Log #%llu: ignoring %s",
(unsigned long long)log, status.ToString().c_str()); (unsigned long long) log,
status.ToString().c_str());
status = Status::OK(); // Keep going with rest of file status = Status::OK(); // Keep going with rest of file
} }
} }
@ -206,14 +227,16 @@ class Repairer {
status = BuildTable(dbname_, env_, options_, table_cache_, iter, &meta); status = BuildTable(dbname_, env_, options_, table_cache_, iter, &meta);
delete iter; delete iter;
mem->Unref(); mem->Unref();
mem = nullptr; mem = NULL;
if (status.ok()) { if (status.ok()) {
if (meta.file_size > 0) { if (meta.file_size > 0) {
table_numbers_.push_back(meta.number); table_numbers_.push_back(meta.number);
} }
} }
Log(options_.info_log, "Log #%llu: %d ops saved to Table #%llu %s", Log(options_.info_log, "Log #%llu: %d ops saved to Table #%llu %s",
(unsigned long long)log, counter, (unsigned long long)meta.number, (unsigned long long) log,
counter,
(unsigned long long) meta.number,
status.ToString().c_str()); status.ToString().c_str());
return status; return status;
} }
@ -249,7 +272,8 @@ class Repairer {
ArchiveFile(TableFileName(dbname_, number)); ArchiveFile(TableFileName(dbname_, number));
ArchiveFile(SSTTableFileName(dbname_, number)); ArchiveFile(SSTTableFileName(dbname_, number));
Log(options_.info_log, "Table #%llu: dropped: %s", Log(options_.info_log, "Table #%llu: dropped: %s",
(unsigned long long)t.meta.number, status.ToString().c_str()); (unsigned long long) t.meta.number,
status.ToString().c_str());
return; return;
} }
@ -263,7 +287,8 @@ class Repairer {
Slice key = iter->key(); Slice key = iter->key();
if (!ParseInternalKey(key, &parsed)) { if (!ParseInternalKey(key, &parsed)) {
Log(options_.info_log, "Table #%llu: unparsable key %s", Log(options_.info_log, "Table #%llu: unparsable key %s",
(unsigned long long)t.meta.number, EscapeString(key).c_str()); (unsigned long long) t.meta.number,
EscapeString(key).c_str());
continue; continue;
} }
@ -282,7 +307,9 @@ class Repairer {
} }
delete iter; delete iter;
Log(options_.info_log, "Table #%llu: %d entries %s", Log(options_.info_log, "Table #%llu: %d entries %s",
(unsigned long long)t.meta.number, counter, status.ToString().c_str()); (unsigned long long) t.meta.number,
counter,
status.ToString().c_str());
if (status.ok()) { if (status.ok()) {
tables_.push_back(t); tables_.push_back(t);
@ -323,25 +350,25 @@ class Repairer {
} }
} }
delete builder; delete builder;
builder = nullptr; builder = NULL;
if (s.ok()) { if (s.ok()) {
s = file->Close(); s = file->Close();
} }
delete file; delete file;
file = nullptr; file = NULL;
if (counter > 0 && s.ok()) { if (counter > 0 && s.ok()) {
std::string orig = TableFileName(dbname_, t.meta.number); std::string orig = TableFileName(dbname_, t.meta.number);
s = env_->RenameFile(copy, orig); s = env_->RenameFile(copy, orig);
if (s.ok()) { if (s.ok()) {
Log(options_.info_log, "Table #%llu: %d entries repaired", Log(options_.info_log, "Table #%llu: %d entries repaired",
(unsigned long long)t.meta.number, counter); (unsigned long long) t.meta.number, counter);
tables_.push_back(t); tables_.push_back(t);
} }
} }
if (!s.ok()) { if (!s.ok()) {
env_->RemoveFile(copy); env_->DeleteFile(copy);
} }
} }
@ -368,12 +395,11 @@ class Repairer {
for (size_t i = 0; i < tables_.size(); i++) { for (size_t i = 0; i < tables_.size(); i++) {
// TODO(opt): separate out into multiple levels // TODO(opt): separate out into multiple levels
const TableInfo& t = tables_[i]; const TableInfo& t = tables_[i];
edit_.AddFile(0, t.meta.number, t.meta.file_size, t.meta.smallest, edit_.AddFile(0, t.meta.number, t.meta.file_size,
t.meta.largest); t.meta.smallest, t.meta.largest);
} }
// std::fprintf(stderr, //fprintf(stderr, "NewDescriptor:\n%s\n", edit_.DebugString().c_str());
// "NewDescriptor:\n%s\n", edit_.DebugString().c_str());
{ {
log::Writer log(file); log::Writer log(file);
std::string record; std::string record;
@ -384,10 +410,10 @@ class Repairer {
status = file->Close(); status = file->Close();
} }
delete file; delete file;
file = nullptr; file = NULL;
if (!status.ok()) { if (!status.ok()) {
env_->RemoveFile(tmp); env_->DeleteFile(tmp);
} else { } else {
// Discard older manifests // Discard older manifests
for (size_t i = 0; i < manifests_.size(); i++) { for (size_t i = 0; i < manifests_.size(); i++) {
@ -399,7 +425,7 @@ class Repairer {
if (status.ok()) { if (status.ok()) {
status = SetCurrentFile(env_, dbname_, 1); status = SetCurrentFile(env_, dbname_, 1);
} else { } else {
env_->RemoveFile(tmp); env_->DeleteFile(tmp);
} }
} }
return status; return status;
@ -412,34 +438,18 @@ class Repairer {
// dir/lost/foo // dir/lost/foo
const char* slash = strrchr(fname.c_str(), '/'); const char* slash = strrchr(fname.c_str(), '/');
std::string new_dir; std::string new_dir;
if (slash != nullptr) { if (slash != NULL) {
new_dir.assign(fname.data(), slash - fname.data()); new_dir.assign(fname.data(), slash - fname.data());
} }
new_dir.append("/lost"); new_dir.append("/lost");
env_->CreateDir(new_dir); // Ignore error env_->CreateDir(new_dir); // Ignore error
std::string new_file = new_dir; std::string new_file = new_dir;
new_file.append("/"); new_file.append("/");
new_file.append((slash == nullptr) ? fname.c_str() : slash + 1); new_file.append((slash == NULL) ? fname.c_str() : slash + 1);
Status s = env_->RenameFile(fname, new_file); Status s = env_->RenameFile(fname, new_file);
Log(options_.info_log, "Archiving %s: %s\n", fname.c_str(), Log(options_.info_log, "Archiving %s: %s\n",
s.ToString().c_str()); fname.c_str(), s.ToString().c_str());
} }
const std::string dbname_;
Env* const env_;
InternalKeyComparator const icmp_;
InternalFilterPolicy const ipolicy_;
const Options options_;
bool owns_info_log_;
bool owns_cache_;
TableCache* table_cache_;
VersionEdit edit_;
std::vector<std::string> manifests_;
std::vector<uint64_t> table_numbers_;
std::vector<uint64_t> logs_;
std::vector<TableInfo> tables_;
uint64_t next_file_number_;
}; };
} // namespace } // namespace

View File

@ -27,16 +27,17 @@
// //
// ... prev vs. next pointer ordering ... // ... prev vs. next pointer ordering ...
#include <atomic> #include <assert.h>
#include <cassert> #include <stdlib.h>
#include <cstdlib> #include "port/port.h"
#include "util/arena.h" #include "util/arena.h"
#include "util/random.h" #include "util/random.h"
namespace leveldb { namespace leveldb {
template <typename Key, class Comparator> class Arena;
template<typename Key, class Comparator>
class SkipList { class SkipList {
private: private:
struct Node; struct Node;
@ -47,9 +48,6 @@ class SkipList {
// must remain allocated for the lifetime of the skiplist object. // must remain allocated for the lifetime of the skiplist object.
explicit SkipList(Comparator cmp, Arena* arena); explicit SkipList(Comparator cmp, Arena* arena);
SkipList(const SkipList&) = delete;
SkipList& operator=(const SkipList&) = delete;
// Insert key into the list. // Insert key into the list.
// REQUIRES: nothing that compares equal to key is currently in the list. // REQUIRES: nothing that compares equal to key is currently in the list.
void Insert(const Key& key); void Insert(const Key& key);
@ -99,10 +97,24 @@ class SkipList {
private: private:
enum { kMaxHeight = 12 }; enum { kMaxHeight = 12 };
// Immutable after construction
Comparator const compare_;
Arena* const arena_; // Arena used for allocations of nodes
Node* const head_;
// Modified only by Insert(). Read racily by readers, but stale
// values are ok.
port::AtomicPointer max_height_; // Height of the entire list
inline int GetMaxHeight() const { inline int GetMaxHeight() const {
return max_height_.load(std::memory_order_relaxed); return static_cast<int>(
reinterpret_cast<intptr_t>(max_height_.NoBarrier_Load()));
} }
// Read/written only by Insert().
Random rnd_;
Node* NewNode(const Key& key, int height); Node* NewNode(const Key& key, int height);
int RandomHeight(); int RandomHeight();
bool Equal(const Key& a, const Key& b) const { return (compare_(a, b) == 0); } bool Equal(const Key& a, const Key& b) const { return (compare_(a, b) == 0); }
@ -111,9 +123,9 @@ class SkipList {
bool KeyIsAfterNode(const Key& key, Node* n) const; bool KeyIsAfterNode(const Key& key, Node* n) const;
// Return the earliest node that comes at or after key. // Return the earliest node that comes at or after key.
// Return nullptr if there is no such node. // Return NULL if there is no such node.
// //
// If prev is non-null, fills prev[level] with pointer to previous // If prev is non-NULL, fills prev[level] with pointer to previous
// node at "level" for every level in [0..max_height_-1]. // node at "level" for every level in [0..max_height_-1].
Node* FindGreaterOrEqual(const Key& key, Node** prev) const; Node* FindGreaterOrEqual(const Key& key, Node** prev) const;
@ -125,24 +137,15 @@ class SkipList {
// Return head_ if list is empty. // Return head_ if list is empty.
Node* FindLast() const; Node* FindLast() const;
// Immutable after construction // No copying allowed
Comparator const compare_; SkipList(const SkipList&);
Arena* const arena_; // Arena used for allocations of nodes void operator=(const SkipList&);
Node* const head_;
// Modified only by Insert(). Read racily by readers, but stale
// values are ok.
std::atomic<int> max_height_; // Height of the entire list
// Read/written only by Insert().
Random rnd_;
}; };
// Implementation details follow // Implementation details follow
template <typename Key, class Comparator> template<typename Key, class Comparator>
struct SkipList<Key, Comparator>::Node { struct SkipList<Key,Comparator>::Node {
explicit Node(const Key& k) : key(k) {} explicit Node(const Key& k) : key(k) { }
Key const key; Key const key;
@ -152,96 +155,96 @@ struct SkipList<Key, Comparator>::Node {
assert(n >= 0); assert(n >= 0);
// Use an 'acquire load' so that we observe a fully initialized // Use an 'acquire load' so that we observe a fully initialized
// version of the returned Node. // version of the returned Node.
return next_[n].load(std::memory_order_acquire); return reinterpret_cast<Node*>(next_[n].Acquire_Load());
} }
void SetNext(int n, Node* x) { void SetNext(int n, Node* x) {
assert(n >= 0); assert(n >= 0);
// Use a 'release store' so that anybody who reads through this // Use a 'release store' so that anybody who reads through this
// pointer observes a fully initialized version of the inserted node. // pointer observes a fully initialized version of the inserted node.
next_[n].store(x, std::memory_order_release); next_[n].Release_Store(x);
} }
// No-barrier variants that can be safely used in a few locations. // No-barrier variants that can be safely used in a few locations.
Node* NoBarrier_Next(int n) { Node* NoBarrier_Next(int n) {
assert(n >= 0); assert(n >= 0);
return next_[n].load(std::memory_order_relaxed); return reinterpret_cast<Node*>(next_[n].NoBarrier_Load());
} }
void NoBarrier_SetNext(int n, Node* x) { void NoBarrier_SetNext(int n, Node* x) {
assert(n >= 0); assert(n >= 0);
next_[n].store(x, std::memory_order_relaxed); next_[n].NoBarrier_Store(x);
} }
private: private:
// Array of length equal to the node height. next_[0] is lowest level link. // Array of length equal to the node height. next_[0] is lowest level link.
std::atomic<Node*> next_[1]; port::AtomicPointer next_[1];
}; };
template <typename Key, class Comparator> template<typename Key, class Comparator>
typename SkipList<Key, Comparator>::Node* SkipList<Key, Comparator>::NewNode( typename SkipList<Key,Comparator>::Node*
const Key& key, int height) { SkipList<Key,Comparator>::NewNode(const Key& key, int height) {
char* const node_memory = arena_->AllocateAligned( char* mem = arena_->AllocateAligned(
sizeof(Node) + sizeof(std::atomic<Node*>) * (height - 1)); sizeof(Node) + sizeof(port::AtomicPointer) * (height - 1));
return new (node_memory) Node(key); return new (mem) Node(key);
} }
template <typename Key, class Comparator> template<typename Key, class Comparator>
inline SkipList<Key, Comparator>::Iterator::Iterator(const SkipList* list) { inline SkipList<Key,Comparator>::Iterator::Iterator(const SkipList* list) {
list_ = list; list_ = list;
node_ = nullptr; node_ = NULL;
} }
template <typename Key, class Comparator> template<typename Key, class Comparator>
inline bool SkipList<Key, Comparator>::Iterator::Valid() const { inline bool SkipList<Key,Comparator>::Iterator::Valid() const {
return node_ != nullptr; return node_ != NULL;
} }
template <typename Key, class Comparator> template<typename Key, class Comparator>
inline const Key& SkipList<Key, Comparator>::Iterator::key() const { inline const Key& SkipList<Key,Comparator>::Iterator::key() const {
assert(Valid()); assert(Valid());
return node_->key; return node_->key;
} }
template <typename Key, class Comparator> template<typename Key, class Comparator>
inline void SkipList<Key, Comparator>::Iterator::Next() { inline void SkipList<Key,Comparator>::Iterator::Next() {
assert(Valid()); assert(Valid());
node_ = node_->Next(0); node_ = node_->Next(0);
} }
template <typename Key, class Comparator> template<typename Key, class Comparator>
inline void SkipList<Key, Comparator>::Iterator::Prev() { inline void SkipList<Key,Comparator>::Iterator::Prev() {
// Instead of using explicit "prev" links, we just search for the // Instead of using explicit "prev" links, we just search for the
// last node that falls before key. // last node that falls before key.
assert(Valid()); assert(Valid());
node_ = list_->FindLessThan(node_->key); node_ = list_->FindLessThan(node_->key);
if (node_ == list_->head_) { if (node_ == list_->head_) {
node_ = nullptr; node_ = NULL;
} }
} }
template <typename Key, class Comparator> template<typename Key, class Comparator>
inline void SkipList<Key, Comparator>::Iterator::Seek(const Key& target) { inline void SkipList<Key,Comparator>::Iterator::Seek(const Key& target) {
node_ = list_->FindGreaterOrEqual(target, nullptr); node_ = list_->FindGreaterOrEqual(target, NULL);
} }
template <typename Key, class Comparator> template<typename Key, class Comparator>
inline void SkipList<Key, Comparator>::Iterator::SeekToFirst() { inline void SkipList<Key,Comparator>::Iterator::SeekToFirst() {
node_ = list_->head_->Next(0); node_ = list_->head_->Next(0);
} }
template <typename Key, class Comparator> template<typename Key, class Comparator>
inline void SkipList<Key, Comparator>::Iterator::SeekToLast() { inline void SkipList<Key,Comparator>::Iterator::SeekToLast() {
node_ = list_->FindLast(); node_ = list_->FindLast();
if (node_ == list_->head_) { if (node_ == list_->head_) {
node_ = nullptr; node_ = NULL;
} }
} }
template <typename Key, class Comparator> template<typename Key, class Comparator>
int SkipList<Key, Comparator>::RandomHeight() { int SkipList<Key,Comparator>::RandomHeight() {
// Increase height with probability 1 in kBranching // Increase height with probability 1 in kBranching
static const unsigned int kBranching = 4; static const unsigned int kBranching = 4;
int height = 1; int height = 1;
while (height < kMaxHeight && rnd_.OneIn(kBranching)) { while (height < kMaxHeight && ((rnd_.Next() % kBranching) == 0)) {
height++; height++;
} }
assert(height > 0); assert(height > 0);
@ -249,16 +252,15 @@ int SkipList<Key, Comparator>::RandomHeight() {
return height; return height;
} }
template <typename Key, class Comparator> template<typename Key, class Comparator>
bool SkipList<Key, Comparator>::KeyIsAfterNode(const Key& key, Node* n) const { bool SkipList<Key,Comparator>::KeyIsAfterNode(const Key& key, Node* n) const {
// null n is considered infinite // NULL n is considered infinite
return (n != nullptr) && (compare_(n->key, key) < 0); return (n != NULL) && (compare_(n->key, key) < 0);
} }
template <typename Key, class Comparator> template<typename Key, class Comparator>
typename SkipList<Key, Comparator>::Node* typename SkipList<Key,Comparator>::Node* SkipList<Key,Comparator>::FindGreaterOrEqual(const Key& key, Node** prev)
SkipList<Key, Comparator>::FindGreaterOrEqual(const Key& key, const {
Node** prev) const {
Node* x = head_; Node* x = head_;
int level = GetMaxHeight() - 1; int level = GetMaxHeight() - 1;
while (true) { while (true) {
@ -267,7 +269,7 @@ SkipList<Key, Comparator>::FindGreaterOrEqual(const Key& key,
// Keep searching in this list // Keep searching in this list
x = next; x = next;
} else { } else {
if (prev != nullptr) prev[level] = x; if (prev != NULL) prev[level] = x;
if (level == 0) { if (level == 0) {
return next; return next;
} else { } else {
@ -278,15 +280,15 @@ SkipList<Key, Comparator>::FindGreaterOrEqual(const Key& key,
} }
} }
template <typename Key, class Comparator> template<typename Key, class Comparator>
typename SkipList<Key, Comparator>::Node* typename SkipList<Key,Comparator>::Node*
SkipList<Key, Comparator>::FindLessThan(const Key& key) const { SkipList<Key,Comparator>::FindLessThan(const Key& key) const {
Node* x = head_; Node* x = head_;
int level = GetMaxHeight() - 1; int level = GetMaxHeight() - 1;
while (true) { while (true) {
assert(x == head_ || compare_(x->key, key) < 0); assert(x == head_ || compare_(x->key, key) < 0);
Node* next = x->Next(level); Node* next = x->Next(level);
if (next == nullptr || compare_(next->key, key) >= 0) { if (next == NULL || compare_(next->key, key) >= 0) {
if (level == 0) { if (level == 0) {
return x; return x;
} else { } else {
@ -299,14 +301,14 @@ SkipList<Key, Comparator>::FindLessThan(const Key& key) const {
} }
} }
template <typename Key, class Comparator> template<typename Key, class Comparator>
typename SkipList<Key, Comparator>::Node* SkipList<Key, Comparator>::FindLast() typename SkipList<Key,Comparator>::Node* SkipList<Key,Comparator>::FindLast()
const { const {
Node* x = head_; Node* x = head_;
int level = GetMaxHeight() - 1; int level = GetMaxHeight() - 1;
while (true) { while (true) {
Node* next = x->Next(level); Node* next = x->Next(level);
if (next == nullptr) { if (next == NULL) {
if (level == 0) { if (level == 0) {
return x; return x;
} else { } else {
@ -319,41 +321,43 @@ typename SkipList<Key, Comparator>::Node* SkipList<Key, Comparator>::FindLast()
} }
} }
template <typename Key, class Comparator> template<typename Key, class Comparator>
SkipList<Key, Comparator>::SkipList(Comparator cmp, Arena* arena) SkipList<Key,Comparator>::SkipList(Comparator cmp, Arena* arena)
: compare_(cmp), : compare_(cmp),
arena_(arena), arena_(arena),
head_(NewNode(0 /* any key will do */, kMaxHeight)), head_(NewNode(0 /* any key will do */, kMaxHeight)),
max_height_(1), max_height_(reinterpret_cast<void*>(1)),
rnd_(0xdeadbeef) { rnd_(0xdeadbeef) {
for (int i = 0; i < kMaxHeight; i++) { for (int i = 0; i < kMaxHeight; i++) {
head_->SetNext(i, nullptr); head_->SetNext(i, NULL);
} }
} }
template <typename Key, class Comparator> template<typename Key, class Comparator>
void SkipList<Key, Comparator>::Insert(const Key& key) { void SkipList<Key,Comparator>::Insert(const Key& key) {
// TODO(opt): We can use a barrier-free variant of FindGreaterOrEqual() // TODO(opt): We can use a barrier-free variant of FindGreaterOrEqual()
// here since Insert() is externally synchronized. // here since Insert() is externally synchronized.
Node* prev[kMaxHeight]; Node* prev[kMaxHeight];
Node* x = FindGreaterOrEqual(key, prev); Node* x = FindGreaterOrEqual(key, prev);
// Our data structure does not allow duplicate insertion // Our data structure does not allow duplicate insertion
assert(x == nullptr || !Equal(key, x->key)); assert(x == NULL || !Equal(key, x->key));
int height = RandomHeight(); int height = RandomHeight();
if (height > GetMaxHeight()) { if (height > GetMaxHeight()) {
for (int i = GetMaxHeight(); i < height; i++) { for (int i = GetMaxHeight(); i < height; i++) {
prev[i] = head_; prev[i] = head_;
} }
//fprintf(stderr, "Change height from %d to %d\n", max_height_, height);
// It is ok to mutate max_height_ without any synchronization // It is ok to mutate max_height_ without any synchronization
// with concurrent readers. A concurrent reader that observes // with concurrent readers. A concurrent reader that observes
// the new value of max_height_ will see either the old value of // the new value of max_height_ will see either the old value of
// new level pointers from head_ (nullptr), or a new value set in // new level pointers from head_ (NULL), or a new value set in
// the loop below. In the former case the reader will // the loop below. In the former case the reader will
// immediately drop to the next level since nullptr sorts after all // immediately drop to the next level since NULL sorts after all
// keys. In the latter case the reader will use the new node. // keys. In the latter case the reader will use the new node.
max_height_.store(height, std::memory_order_relaxed); max_height_.NoBarrier_Store(reinterpret_cast<void*>(height));
} }
x = NewNode(key, height); x = NewNode(key, height);
@ -365,10 +369,10 @@ void SkipList<Key, Comparator>::Insert(const Key& key) {
} }
} }
template <typename Key, class Comparator> template<typename Key, class Comparator>
bool SkipList<Key, Comparator>::Contains(const Key& key) const { bool SkipList<Key,Comparator>::Contains(const Key& key) const {
Node* x = FindGreaterOrEqual(key, nullptr); Node* x = FindGreaterOrEqual(key, NULL);
if (x != nullptr && Equal(key, x->key)) { if (x != NULL && Equal(key, x->key)) {
return true; return true;
} else { } else {
return false; return false;

View File

@ -3,18 +3,12 @@
// found in the LICENSE file. See the AUTHORS file for names of contributors. // found in the LICENSE file. See the AUTHORS file for names of contributors.
#include "db/skiplist.h" #include "db/skiplist.h"
#include <atomic>
#include <set> #include <set>
#include "gtest/gtest.h"
#include "leveldb/env.h" #include "leveldb/env.h"
#include "port/port.h"
#include "port/thread_annotations.h"
#include "util/arena.h" #include "util/arena.h"
#include "util/hash.h" #include "util/hash.h"
#include "util/random.h" #include "util/random.h"
#include "util/testutil.h" #include "util/testharness.h"
namespace leveldb { namespace leveldb {
@ -32,6 +26,8 @@ struct Comparator {
} }
}; };
class SkipTest { };
TEST(SkipTest, Empty) { TEST(SkipTest, Empty) {
Arena arena; Arena arena;
Comparator cmp; Comparator cmp;
@ -116,7 +112,8 @@ TEST(SkipTest, InsertAndLookup) {
// Compare against model iterator // Compare against model iterator
for (std::set<Key>::reverse_iterator model_iter = keys.rbegin(); for (std::set<Key>::reverse_iterator model_iter = keys.rbegin();
model_iter != keys.rend(); ++model_iter) { model_iter != keys.rend();
++model_iter) {
ASSERT_TRUE(iter.Valid()); ASSERT_TRUE(iter.Valid());
ASSERT_EQ(*model_iter, iter.key()); ASSERT_EQ(*model_iter, iter.key());
iter.Prev(); iter.Prev();
@ -129,7 +126,7 @@ TEST(SkipTest, InsertAndLookup) {
// concurrent readers (with no synchronization other than when a // concurrent readers (with no synchronization other than when a
// reader's iterator is created), the reader always observes all the // reader's iterator is created), the reader always observes all the
// data that was present in the skip list when the iterator was // data that was present in the skip list when the iterator was
// constructed. Because insertions are happening concurrently, we may // constructor. Because insertions are happening concurrently, we may
// also observe new values that were inserted since the iterator was // also observe new values that were inserted since the iterator was
// constructed, but we should never miss any values that were present // constructed, but we should never miss any values that were present
// at iterator construction time. // at iterator construction time.
@ -151,19 +148,19 @@ TEST(SkipTest, InsertAndLookup) {
// been concurrently added since the iterator started. // been concurrently added since the iterator started.
class ConcurrentTest { class ConcurrentTest {
private: private:
static constexpr uint32_t K = 4; static const uint32_t K = 4;
static uint64_t key(Key key) { return (key >> 40); } static uint64_t key(Key key) { return (key >> 40); }
static uint64_t gen(Key key) { return (key >> 8) & 0xffffffffu; } static uint64_t gen(Key key) { return (key >> 8) & 0xffffffffu; }
static uint64_t hash(Key key) { return key & 0xff; } static uint64_t hash(Key key) { return key & 0xff; }
static uint64_t HashNumbers(uint64_t k, uint64_t g) { static uint64_t HashNumbers(uint64_t k, uint64_t g) {
uint64_t data[2] = {k, g}; uint64_t data[2] = { k, g };
return Hash(reinterpret_cast<char*>(data), sizeof(data), 0); return Hash(reinterpret_cast<char*>(data), sizeof(data), 0);
} }
static Key MakeKey(uint64_t k, uint64_t g) { static Key MakeKey(uint64_t k, uint64_t g) {
static_assert(sizeof(Key) == sizeof(uint64_t), ""); assert(sizeof(Key) == sizeof(uint64_t));
assert(k <= K); // We sometimes pass K to seek to the end of the skiplist assert(k <= K); // We sometimes pass K to seek to the end of the skiplist
assert(g <= 0xffffffffu); assert(g <= 0xffffffffu);
return ((k << 40) | (g << 8) | (HashNumbers(k, g) & 0xff)); return ((k << 40) | (g << 8) | (HashNumbers(k, g) & 0xff));
@ -189,11 +186,13 @@ class ConcurrentTest {
// Per-key generation // Per-key generation
struct State { struct State {
std::atomic<int> generation[K]; port::AtomicPointer generation[K];
void Set(int k, int v) { void Set(int k, intptr_t v) {
generation[k].store(v, std::memory_order_release); generation[k].Release_Store(reinterpret_cast<void*>(v));
}
intptr_t Get(int k) {
return reinterpret_cast<intptr_t>(generation[k].Acquire_Load());
} }
int Get(int k) { return generation[k].load(std::memory_order_acquire); }
State() { State() {
for (int k = 0; k < K; k++) { for (int k = 0; k < K; k++) {
@ -212,7 +211,7 @@ class ConcurrentTest {
SkipList<Key, Comparator> list_; SkipList<Key, Comparator> list_;
public: public:
ConcurrentTest() : list_(Comparator(), &arena_) {} ConcurrentTest() : list_(Comparator(), &arena_) { }
// REQUIRES: External synchronization // REQUIRES: External synchronization
void WriteStep(Random* rnd) { void WriteStep(Random* rnd) {
@ -251,9 +250,11 @@ class ConcurrentTest {
// Note that generation 0 is never inserted, so it is ok if // Note that generation 0 is never inserted, so it is ok if
// <*,0,*> is missing. // <*,0,*> is missing.
ASSERT_TRUE((gen(pos) == 0) || ASSERT_TRUE((gen(pos) == 0) ||
(gen(pos) > static_cast<Key>(initial_state.Get(key(pos))))) (gen(pos) > initial_state.Get(key(pos)))
<< "key: " << key(pos) << "; gen: " << gen(pos) ) << "key: " << key(pos)
<< "; initgen: " << initial_state.Get(key(pos)); << "; gen: " << gen(pos)
<< "; initgen: "
<< initial_state.Get(key(pos));
// Advance to next key in the valid key space // Advance to next key in the valid key space
if (key(pos) < key(current)) { if (key(pos) < key(current)) {
@ -280,9 +281,7 @@ class ConcurrentTest {
} }
} }
}; };
const uint32_t ConcurrentTest::K;
// Needed when building in C++11 mode.
constexpr uint32_t ConcurrentTest::K;
// Simple test that does single-threaded testing of the ConcurrentTest // Simple test that does single-threaded testing of the ConcurrentTest
// scaffolding. // scaffolding.
@ -299,14 +298,21 @@ class TestState {
public: public:
ConcurrentTest t_; ConcurrentTest t_;
int seed_; int seed_;
std::atomic<bool> quit_flag_; port::AtomicPointer quit_flag_;
enum ReaderState { STARTING, RUNNING, DONE }; enum ReaderState {
STARTING,
RUNNING,
DONE
};
explicit TestState(int s) explicit TestState(int s)
: seed_(s), quit_flag_(false), state_(STARTING), state_cv_(&mu_) {} : seed_(s),
quit_flag_(NULL),
state_(STARTING),
state_cv_(&mu_) {}
void Wait(ReaderState s) LOCKS_EXCLUDED(mu_) { void Wait(ReaderState s) {
mu_.Lock(); mu_.Lock();
while (state_ != s) { while (state_ != s) {
state_cv_.Wait(); state_cv_.Wait();
@ -314,7 +320,7 @@ class TestState {
mu_.Unlock(); mu_.Unlock();
} }
void Change(ReaderState s) LOCKS_EXCLUDED(mu_) { void Change(ReaderState s) {
mu_.Lock(); mu_.Lock();
state_ = s; state_ = s;
state_cv_.Signal(); state_cv_.Signal();
@ -323,8 +329,8 @@ class TestState {
private: private:
port::Mutex mu_; port::Mutex mu_;
ReaderState state_ GUARDED_BY(mu_); ReaderState state_;
port::CondVar state_cv_ GUARDED_BY(mu_); port::CondVar state_cv_;
}; };
static void ConcurrentReader(void* arg) { static void ConcurrentReader(void* arg) {
@ -332,7 +338,7 @@ static void ConcurrentReader(void* arg) {
Random rnd(state->seed_); Random rnd(state->seed_);
int64_t reads = 0; int64_t reads = 0;
state->Change(TestState::RUNNING); state->Change(TestState::RUNNING);
while (!state->quit_flag_.load(std::memory_order_acquire)) { while (!state->quit_flag_.Acquire_Load()) {
state->t_.ReadStep(&rnd); state->t_.ReadStep(&rnd);
++reads; ++reads;
} }
@ -346,7 +352,7 @@ static void RunConcurrent(int run) {
const int kSize = 1000; const int kSize = 1000;
for (int i = 0; i < N; i++) { for (int i = 0; i < N; i++) {
if ((i % 100) == 0) { if ((i % 100) == 0) {
std::fprintf(stderr, "Run %d of %d\n", i, N); fprintf(stderr, "Run %d of %d\n", i, N);
} }
TestState state(seed + 1); TestState state(seed + 1);
Env::Default()->Schedule(ConcurrentReader, &state); Env::Default()->Schedule(ConcurrentReader, &state);
@ -354,7 +360,7 @@ static void RunConcurrent(int run) {
for (int i = 0; i < kSize; i++) { for (int i = 0; i < kSize; i++) {
state.t_.WriteStep(&rnd); state.t_.WriteStep(&rnd);
} }
state.quit_flag_.store(true, std::memory_order_release); state.quit_flag_.Release_Store(&state); // Any non-NULL arg will do
state.Wait(TestState::DONE); state.Wait(TestState::DONE);
} }
} }
@ -366,3 +372,7 @@ TEST(SkipTest, Concurrent4) { RunConcurrent(4); }
TEST(SkipTest, Concurrent5) { RunConcurrent(5); } TEST(SkipTest, Concurrent5) { RunConcurrent(5); }
} // namespace leveldb } // namespace leveldb
int main(int argc, char** argv) {
return leveldb::test::RunAllTests();
}

View File

@ -16,78 +16,50 @@ class SnapshotList;
// Each SnapshotImpl corresponds to a particular sequence number. // Each SnapshotImpl corresponds to a particular sequence number.
class SnapshotImpl : public Snapshot { class SnapshotImpl : public Snapshot {
public: public:
SnapshotImpl(SequenceNumber sequence_number) SequenceNumber number_; // const after creation
: sequence_number_(sequence_number) {}
SequenceNumber sequence_number() const { return sequence_number_; }
private: private:
friend class SnapshotList; friend class SnapshotList;
// SnapshotImpl is kept in a doubly-linked circular list. The SnapshotList // SnapshotImpl is kept in a doubly-linked circular list
// implementation operates on the next/previous fields directly.
SnapshotImpl* prev_; SnapshotImpl* prev_;
SnapshotImpl* next_; SnapshotImpl* next_;
const SequenceNumber sequence_number_; SnapshotList* list_; // just for sanity checks
#if !defined(NDEBUG)
SnapshotList* list_ = nullptr;
#endif // !defined(NDEBUG)
}; };
class SnapshotList { class SnapshotList {
public: public:
SnapshotList() : head_(0) { SnapshotList() {
head_.prev_ = &head_; list_.prev_ = &list_;
head_.next_ = &head_; list_.next_ = &list_;
} }
bool empty() const { return head_.next_ == &head_; } bool empty() const { return list_.next_ == &list_; }
SnapshotImpl* oldest() const { SnapshotImpl* oldest() const { assert(!empty()); return list_.next_; }
assert(!empty()); SnapshotImpl* newest() const { assert(!empty()); return list_.prev_; }
return head_.next_;
} const SnapshotImpl* New(SequenceNumber seq) {
SnapshotImpl* newest() const { SnapshotImpl* s = new SnapshotImpl;
assert(!empty()); s->number_ = seq;
return head_.prev_; s->list_ = this;
s->next_ = &list_;
s->prev_ = list_.prev_;
s->prev_->next_ = s;
s->next_->prev_ = s;
return s;
} }
// Creates a SnapshotImpl and appends it to the end of the list. void Delete(const SnapshotImpl* s) {
SnapshotImpl* New(SequenceNumber sequence_number) { assert(s->list_ == this);
assert(empty() || newest()->sequence_number_ <= sequence_number); s->prev_->next_ = s->next_;
s->next_->prev_ = s->prev_;
SnapshotImpl* snapshot = new SnapshotImpl(sequence_number); delete s;
#if !defined(NDEBUG)
snapshot->list_ = this;
#endif // !defined(NDEBUG)
snapshot->next_ = &head_;
snapshot->prev_ = head_.prev_;
snapshot->prev_->next_ = snapshot;
snapshot->next_->prev_ = snapshot;
return snapshot;
}
// Removes a SnapshotImpl from this list.
//
// The snapshot must have been created by calling New() on this list.
//
// The snapshot pointer should not be const, because its memory is
// deallocated. However, that would force us to change DB::ReleaseSnapshot(),
// which is in the API, and currently takes a const Snapshot.
void Delete(const SnapshotImpl* snapshot) {
#if !defined(NDEBUG)
assert(snapshot->list_ == this);
#endif // !defined(NDEBUG)
snapshot->prev_->next_ = snapshot->next_;
snapshot->next_->prev_ = snapshot->prev_;
delete snapshot;
} }
private: private:
// Dummy head of doubly-linked list of snapshots // Dummy head of doubly-linked list of snapshots
SnapshotImpl head_; SnapshotImpl list_;
}; };
} // namespace leveldb } // namespace leveldb

View File

@ -29,14 +29,18 @@ static void UnrefEntry(void* arg1, void* arg2) {
cache->Release(h); cache->Release(h);
} }
TableCache::TableCache(const std::string& dbname, const Options& options, TableCache::TableCache(const std::string& dbname,
const Options* options,
int entries) int entries)
: env_(options.env), : env_(options->env),
dbname_(dbname), dbname_(dbname),
options_(options), options_(options),
cache_(NewLRUCache(entries)) {} cache_(NewLRUCache(entries)) {
}
TableCache::~TableCache() { delete cache_; } TableCache::~TableCache() {
delete cache_;
}
Status TableCache::FindTable(uint64_t file_number, uint64_t file_size, Status TableCache::FindTable(uint64_t file_number, uint64_t file_size,
Cache::Handle** handle) { Cache::Handle** handle) {
@ -45,10 +49,10 @@ Status TableCache::FindTable(uint64_t file_number, uint64_t file_size,
EncodeFixed64(buf, file_number); EncodeFixed64(buf, file_number);
Slice key(buf, sizeof(buf)); Slice key(buf, sizeof(buf));
*handle = cache_->Lookup(key); *handle = cache_->Lookup(key);
if (*handle == nullptr) { if (*handle == NULL) {
std::string fname = TableFileName(dbname_, file_number); std::string fname = TableFileName(dbname_, file_number);
RandomAccessFile* file = nullptr; RandomAccessFile* file = NULL;
Table* table = nullptr; Table* table = NULL;
s = env_->NewRandomAccessFile(fname, &file); s = env_->NewRandomAccessFile(fname, &file);
if (!s.ok()) { if (!s.ok()) {
std::string old_fname = SSTTableFileName(dbname_, file_number); std::string old_fname = SSTTableFileName(dbname_, file_number);
@ -57,11 +61,11 @@ Status TableCache::FindTable(uint64_t file_number, uint64_t file_size,
} }
} }
if (s.ok()) { if (s.ok()) {
s = Table::Open(options_, file, file_size, &table); s = Table::Open(*options_, file, file_size, &table);
} }
if (!s.ok()) { if (!s.ok()) {
assert(table == nullptr); assert(table == NULL);
delete file; delete file;
// We do not cache error results so that if the error is transient, // We do not cache error results so that if the error is transient,
// or somebody repairs the file, we recover automatically. // or somebody repairs the file, we recover automatically.
@ -76,13 +80,14 @@ Status TableCache::FindTable(uint64_t file_number, uint64_t file_size,
} }
Iterator* TableCache::NewIterator(const ReadOptions& options, Iterator* TableCache::NewIterator(const ReadOptions& options,
uint64_t file_number, uint64_t file_size, uint64_t file_number,
uint64_t file_size,
Table** tableptr) { Table** tableptr) {
if (tableptr != nullptr) { if (tableptr != NULL) {
*tableptr = nullptr; *tableptr = NULL;
} }
Cache::Handle* handle = nullptr; Cache::Handle* handle = NULL;
Status s = FindTable(file_number, file_size, &handle); Status s = FindTable(file_number, file_size, &handle);
if (!s.ok()) { if (!s.ok()) {
return NewErrorIterator(s); return NewErrorIterator(s);
@ -91,21 +96,23 @@ Iterator* TableCache::NewIterator(const ReadOptions& options,
Table* table = reinterpret_cast<TableAndFile*>(cache_->Value(handle))->table; Table* table = reinterpret_cast<TableAndFile*>(cache_->Value(handle))->table;
Iterator* result = table->NewIterator(options); Iterator* result = table->NewIterator(options);
result->RegisterCleanup(&UnrefEntry, cache_, handle); result->RegisterCleanup(&UnrefEntry, cache_, handle);
if (tableptr != nullptr) { if (tableptr != NULL) {
*tableptr = table; *tableptr = table;
} }
return result; return result;
} }
Status TableCache::Get(const ReadOptions& options, uint64_t file_number, Status TableCache::Get(const ReadOptions& options,
uint64_t file_size, const Slice& k, void* arg, uint64_t file_number,
void (*handle_result)(void*, const Slice&, uint64_t file_size,
const Slice&)) { const Slice& k,
Cache::Handle* handle = nullptr; void* arg,
void (*saver)(void*, const Slice&, const Slice&)) {
Cache::Handle* handle = NULL;
Status s = FindTable(file_number, file_size, &handle); Status s = FindTable(file_number, file_size, &handle);
if (s.ok()) { if (s.ok()) {
Table* t = reinterpret_cast<TableAndFile*>(cache_->Value(handle))->table; Table* t = reinterpret_cast<TableAndFile*>(cache_->Value(handle))->table;
s = t->InternalGet(options, k, arg, handle_result); s = t->InternalGet(options, k, arg, saver);
cache_->Release(handle); cache_->Release(handle);
} }
return s; return s;

View File

@ -7,9 +7,8 @@
#ifndef STORAGE_LEVELDB_DB_TABLE_CACHE_H_ #ifndef STORAGE_LEVELDB_DB_TABLE_CACHE_H_
#define STORAGE_LEVELDB_DB_TABLE_CACHE_H_ #define STORAGE_LEVELDB_DB_TABLE_CACHE_H_
#include <cstdint>
#include <string> #include <string>
#include <stdint.h>
#include "db/dbformat.h" #include "db/dbformat.h"
#include "leveldb/cache.h" #include "leveldb/cache.h"
#include "leveldb/table.h" #include "leveldb/table.h"
@ -21,39 +20,40 @@ class Env;
class TableCache { class TableCache {
public: public:
TableCache(const std::string& dbname, const Options& options, int entries); TableCache(const std::string& dbname, const Options* options, int entries);
TableCache(const TableCache&) = delete;
TableCache& operator=(const TableCache&) = delete;
~TableCache(); ~TableCache();
// Return an iterator for the specified file number (the corresponding // Return an iterator for the specified file number (the corresponding
// file length must be exactly "file_size" bytes). If "tableptr" is // file length must be exactly "file_size" bytes). If "tableptr" is
// non-null, also sets "*tableptr" to point to the Table object // non-NULL, also sets "*tableptr" to point to the Table object
// underlying the returned iterator, or to nullptr if no Table object // underlying the returned iterator, or NULL if no Table object underlies
// underlies the returned iterator. The returned "*tableptr" object is owned // the returned iterator. The returned "*tableptr" object is owned by
// by the cache and should not be deleted, and is valid for as long as the // the cache and should not be deleted, and is valid for as long as the
// returned iterator is live. // returned iterator is live.
Iterator* NewIterator(const ReadOptions& options, uint64_t file_number, Iterator* NewIterator(const ReadOptions& options,
uint64_t file_size, Table** tableptr = nullptr); uint64_t file_number,
uint64_t file_size,
Table** tableptr = NULL);
// If a seek to internal key "k" in specified file finds an entry, // If a seek to internal key "k" in specified file finds an entry,
// call (*handle_result)(arg, found_key, found_value). // call (*handle_result)(arg, found_key, found_value).
Status Get(const ReadOptions& options, uint64_t file_number, Status Get(const ReadOptions& options,
uint64_t file_size, const Slice& k, void* arg, uint64_t file_number,
uint64_t file_size,
const Slice& k,
void* arg,
void (*handle_result)(void*, const Slice&, const Slice&)); void (*handle_result)(void*, const Slice&, const Slice&));
// Evict any entry for the specified file number // Evict any entry for the specified file number
void Evict(uint64_t file_number); void Evict(uint64_t file_number);
private: private:
Status FindTable(uint64_t file_number, uint64_t file_size, Cache::Handle**);
Env* const env_; Env* const env_;
const std::string dbname_; const std::string dbname_;
const Options& options_; const Options* options_;
Cache* cache_; Cache* cache_;
Status FindTable(uint64_t file_number, uint64_t file_size, Cache::Handle**);
}; };
} // namespace leveldb } // namespace leveldb

View File

@ -12,15 +12,15 @@ namespace leveldb {
// Tag numbers for serialized VersionEdit. These numbers are written to // Tag numbers for serialized VersionEdit. These numbers are written to
// disk and should not be changed. // disk and should not be changed.
enum Tag { enum Tag {
kComparator = 1, kComparator = 1,
kLogNumber = 2, kLogNumber = 2,
kNextFileNumber = 3, kNextFileNumber = 3,
kLastSequence = 4, kLastSequence = 4,
kCompactPointer = 5, kCompactPointer = 5,
kDeletedFile = 6, kDeletedFile = 6,
kNewFile = 7, kNewFile = 7,
// 8 was used for large value refs // 8 was used for large value refs
kPrevLogNumber = 9 kPrevLogNumber = 9
}; };
void VersionEdit::Clear() { void VersionEdit::Clear() {
@ -34,7 +34,6 @@ void VersionEdit::Clear() {
has_prev_log_number_ = false; has_prev_log_number_ = false;
has_next_file_number_ = false; has_next_file_number_ = false;
has_last_sequence_ = false; has_last_sequence_ = false;
compact_pointers_.clear();
deleted_files_.clear(); deleted_files_.clear();
new_files_.clear(); new_files_.clear();
} }
@ -67,10 +66,12 @@ void VersionEdit::EncodeTo(std::string* dst) const {
PutLengthPrefixedSlice(dst, compact_pointers_[i].second.Encode()); PutLengthPrefixedSlice(dst, compact_pointers_[i].second.Encode());
} }
for (const auto& deleted_file_kvp : deleted_files_) { for (DeletedFileSet::const_iterator iter = deleted_files_.begin();
iter != deleted_files_.end();
++iter) {
PutVarint32(dst, kDeletedFile); PutVarint32(dst, kDeletedFile);
PutVarint32(dst, deleted_file_kvp.first); // level PutVarint32(dst, iter->first); // level
PutVarint64(dst, deleted_file_kvp.second); // file number PutVarint64(dst, iter->second); // file number
} }
for (size_t i = 0; i < new_files_.size(); i++) { for (size_t i = 0; i < new_files_.size(); i++) {
@ -87,7 +88,8 @@ void VersionEdit::EncodeTo(std::string* dst) const {
static bool GetInternalKey(Slice* input, InternalKey* dst) { static bool GetInternalKey(Slice* input, InternalKey* dst) {
Slice str; Slice str;
if (GetLengthPrefixedSlice(input, &str)) { if (GetLengthPrefixedSlice(input, &str)) {
return dst->DecodeFrom(str); dst->DecodeFrom(str);
return true;
} else { } else {
return false; return false;
} }
@ -95,7 +97,8 @@ static bool GetInternalKey(Slice* input, InternalKey* dst) {
static bool GetLevel(Slice* input, int* level) { static bool GetLevel(Slice* input, int* level) {
uint32_t v; uint32_t v;
if (GetVarint32(input, &v) && v < config::kNumLevels) { if (GetVarint32(input, &v) &&
v < config::kNumLevels) {
*level = v; *level = v;
return true; return true;
} else { } else {
@ -106,7 +109,7 @@ static bool GetLevel(Slice* input, int* level) {
Status VersionEdit::DecodeFrom(const Slice& src) { Status VersionEdit::DecodeFrom(const Slice& src) {
Clear(); Clear();
Slice input = src; Slice input = src;
const char* msg = nullptr; const char* msg = NULL;
uint32_t tag; uint32_t tag;
// Temporary storage for parsing // Temporary storage for parsing
@ -116,7 +119,7 @@ Status VersionEdit::DecodeFrom(const Slice& src) {
Slice str; Slice str;
InternalKey key; InternalKey key;
while (msg == nullptr && GetVarint32(&input, &tag)) { while (msg == NULL && GetVarint32(&input, &tag)) {
switch (tag) { switch (tag) {
case kComparator: case kComparator:
if (GetLengthPrefixedSlice(&input, &str)) { if (GetLengthPrefixedSlice(&input, &str)) {
@ -160,7 +163,8 @@ Status VersionEdit::DecodeFrom(const Slice& src) {
break; break;
case kCompactPointer: case kCompactPointer:
if (GetLevel(&input, &level) && GetInternalKey(&input, &key)) { if (GetLevel(&input, &level) &&
GetInternalKey(&input, &key)) {
compact_pointers_.push_back(std::make_pair(level, key)); compact_pointers_.push_back(std::make_pair(level, key));
} else { } else {
msg = "compaction pointer"; msg = "compaction pointer";
@ -168,7 +172,8 @@ Status VersionEdit::DecodeFrom(const Slice& src) {
break; break;
case kDeletedFile: case kDeletedFile:
if (GetLevel(&input, &level) && GetVarint64(&input, &number)) { if (GetLevel(&input, &level) &&
GetVarint64(&input, &number)) {
deleted_files_.insert(std::make_pair(level, number)); deleted_files_.insert(std::make_pair(level, number));
} else { } else {
msg = "deleted file"; msg = "deleted file";
@ -176,7 +181,8 @@ Status VersionEdit::DecodeFrom(const Slice& src) {
break; break;
case kNewFile: case kNewFile:
if (GetLevel(&input, &level) && GetVarint64(&input, &f.number) && if (GetLevel(&input, &level) &&
GetVarint64(&input, &f.number) &&
GetVarint64(&input, &f.file_size) && GetVarint64(&input, &f.file_size) &&
GetInternalKey(&input, &f.smallest) && GetInternalKey(&input, &f.smallest) &&
GetInternalKey(&input, &f.largest)) { GetInternalKey(&input, &f.largest)) {
@ -192,12 +198,12 @@ Status VersionEdit::DecodeFrom(const Slice& src) {
} }
} }
if (msg == nullptr && !input.empty()) { if (msg == NULL && !input.empty()) {
msg = "invalid tag"; msg = "invalid tag";
} }
Status result; Status result;
if (msg != nullptr) { if (msg != NULL) {
result = Status::Corruption("VersionEdit", msg); result = Status::Corruption("VersionEdit", msg);
} }
return result; return result;
@ -232,11 +238,13 @@ std::string VersionEdit::DebugString() const {
r.append(" "); r.append(" ");
r.append(compact_pointers_[i].second.DebugString()); r.append(compact_pointers_[i].second.DebugString());
} }
for (const auto& deleted_files_kvp : deleted_files_) { for (DeletedFileSet::const_iterator iter = deleted_files_.begin();
r.append("\n RemoveFile: "); iter != deleted_files_.end();
AppendNumberTo(&r, deleted_files_kvp.first); ++iter) {
r.append("\n DeleteFile: ");
AppendNumberTo(&r, iter->first);
r.append(" "); r.append(" ");
AppendNumberTo(&r, deleted_files_kvp.second); AppendNumberTo(&r, iter->second);
} }
for (size_t i = 0; i < new_files_.size(); i++) { for (size_t i = 0; i < new_files_.size(); i++) {
const FileMetaData& f = new_files_[i].second; const FileMetaData& f = new_files_[i].second;

View File

@ -8,7 +8,6 @@
#include <set> #include <set>
#include <utility> #include <utility>
#include <vector> #include <vector>
#include "db/dbformat.h" #include "db/dbformat.h"
namespace leveldb { namespace leveldb {
@ -16,20 +15,20 @@ namespace leveldb {
class VersionSet; class VersionSet;
struct FileMetaData { struct FileMetaData {
FileMetaData() : refs(0), allowed_seeks(1 << 30), file_size(0) {}
int refs; int refs;
int allowed_seeks; // Seeks allowed until compaction int allowed_seeks; // Seeks allowed until compaction
uint64_t number; uint64_t number;
uint64_t file_size; // File size in bytes uint64_t file_size; // File size in bytes
InternalKey smallest; // Smallest internal key served by table InternalKey smallest; // Smallest internal key served by table
InternalKey largest; // Largest internal key served by table InternalKey largest; // Largest internal key served by table
FileMetaData() : refs(0), allowed_seeks(1 << 30), file_size(0) { }
}; };
class VersionEdit { class VersionEdit {
public: public:
VersionEdit() { Clear(); } VersionEdit() { Clear(); }
~VersionEdit() = default; ~VersionEdit() { }
void Clear(); void Clear();
@ -60,8 +59,10 @@ class VersionEdit {
// Add the specified file at the specified number. // Add the specified file at the specified number.
// REQUIRES: This version has not been saved (see VersionSet::SaveTo) // REQUIRES: This version has not been saved (see VersionSet::SaveTo)
// REQUIRES: "smallest" and "largest" are smallest and largest keys in file // REQUIRES: "smallest" and "largest" are smallest and largest keys in file
void AddFile(int level, uint64_t file, uint64_t file_size, void AddFile(int level, uint64_t file,
const InternalKey& smallest, const InternalKey& largest) { uint64_t file_size,
const InternalKey& smallest,
const InternalKey& largest) {
FileMetaData f; FileMetaData f;
f.number = file; f.number = file;
f.file_size = file_size; f.file_size = file_size;
@ -71,7 +72,7 @@ class VersionEdit {
} }
// Delete the specified "file" from the specified "level". // Delete the specified "file" from the specified "level".
void RemoveFile(int level, uint64_t file) { void DeleteFile(int level, uint64_t file) {
deleted_files_.insert(std::make_pair(level, file)); deleted_files_.insert(std::make_pair(level, file));
} }
@ -83,7 +84,7 @@ class VersionEdit {
private: private:
friend class VersionSet; friend class VersionSet;
typedef std::set<std::pair<int, uint64_t>> DeletedFileSet; typedef std::set< std::pair<int, uint64_t> > DeletedFileSet;
std::string comparator_; std::string comparator_;
uint64_t log_number_; uint64_t log_number_;
@ -96,9 +97,9 @@ class VersionEdit {
bool has_next_file_number_; bool has_next_file_number_;
bool has_last_sequence_; bool has_last_sequence_;
std::vector<std::pair<int, InternalKey>> compact_pointers_; std::vector< std::pair<int, InternalKey> > compact_pointers_;
DeletedFileSet deleted_files_; DeletedFileSet deleted_files_;
std::vector<std::pair<int, FileMetaData>> new_files_; std::vector< std::pair<int, FileMetaData> > new_files_;
}; };
} // namespace leveldb } // namespace leveldb

View File

@ -3,8 +3,7 @@
// found in the LICENSE file. See the AUTHORS file for names of contributors. // found in the LICENSE file. See the AUTHORS file for names of contributors.
#include "db/version_edit.h" #include "db/version_edit.h"
#include "util/testharness.h"
#include "gtest/gtest.h"
namespace leveldb { namespace leveldb {
@ -18,6 +17,8 @@ static void TestEncodeDecode(const VersionEdit& edit) {
ASSERT_EQ(encoded, encoded2); ASSERT_EQ(encoded, encoded2);
} }
class VersionEditTest { };
TEST(VersionEditTest, EncodeDecode) { TEST(VersionEditTest, EncodeDecode) {
static const uint64_t kBig = 1ull << 50; static const uint64_t kBig = 1ull << 50;
@ -27,7 +28,7 @@ TEST(VersionEditTest, EncodeDecode) {
edit.AddFile(3, kBig + 300 + i, kBig + 400 + i, edit.AddFile(3, kBig + 300 + i, kBig + 400 + i,
InternalKey("foo", kBig + 500 + i, kTypeValue), InternalKey("foo", kBig + 500 + i, kTypeValue),
InternalKey("zoo", kBig + 600 + i, kTypeDeletion)); InternalKey("zoo", kBig + 600 + i, kTypeDeletion));
edit.RemoveFile(4, kBig + 700 + i); edit.DeleteFile(4, kBig + 700 + i);
edit.SetCompactPointer(i, InternalKey("x", kBig + 900 + i, kTypeValue)); edit.SetCompactPointer(i, InternalKey("x", kBig + 900 + i, kTypeValue));
} }
@ -39,3 +40,7 @@ TEST(VersionEditTest, EncodeDecode) {
} }
} // namespace leveldb } // namespace leveldb
int main(int argc, char** argv) {
return leveldb::test::RunAllTests();
}

File diff suppressed because it is too large Load Diff

View File

@ -18,7 +18,6 @@
#include <map> #include <map>
#include <set> #include <set>
#include <vector> #include <vector>
#include "db/dbformat.h" #include "db/dbformat.h"
#include "db/version_edit.h" #include "db/version_edit.h"
#include "port/port.h" #include "port/port.h"
@ -26,9 +25,7 @@
namespace leveldb { namespace leveldb {
namespace log { namespace log { class Writer; }
class Writer;
}
class Compaction; class Compaction;
class Iterator; class Iterator;
@ -42,28 +39,25 @@ class WritableFile;
// Return the smallest index i such that files[i]->largest >= key. // Return the smallest index i such that files[i]->largest >= key.
// Return files.size() if there is no such file. // Return files.size() if there is no such file.
// REQUIRES: "files" contains a sorted list of non-overlapping files. // REQUIRES: "files" contains a sorted list of non-overlapping files.
int FindFile(const InternalKeyComparator& icmp, extern int FindFile(const InternalKeyComparator& icmp,
const std::vector<FileMetaData*>& files, const Slice& key); const std::vector<FileMetaData*>& files,
const Slice& key);
// Returns true iff some file in "files" overlaps the user key range // Returns true iff some file in "files" overlaps the user key range
// [*smallest,*largest]. // [*smallest,*largest].
// smallest==nullptr represents a key smaller than all keys in the DB. // smallest==NULL represents a key smaller than all keys in the DB.
// largest==nullptr represents a key largest than all keys in the DB. // largest==NULL represents a key largest than all keys in the DB.
// REQUIRES: If disjoint_sorted_files, files[] contains disjoint ranges // REQUIRES: If disjoint_sorted_files, files[] contains disjoint ranges
// in sorted order. // in sorted order.
bool SomeFileOverlapsRange(const InternalKeyComparator& icmp, extern bool SomeFileOverlapsRange(
bool disjoint_sorted_files, const InternalKeyComparator& icmp,
const std::vector<FileMetaData*>& files, bool disjoint_sorted_files,
const Slice* smallest_user_key, const std::vector<FileMetaData*>& files,
const Slice* largest_user_key); const Slice* smallest_user_key,
const Slice* largest_user_key);
class Version { class Version {
public: public:
struct GetStats {
FileMetaData* seek_file;
int seek_file_level;
};
// Append to *iters a sequence of iterators that will // Append to *iters a sequence of iterators that will
// yield the contents of this Version when merged together. // yield the contents of this Version when merged together.
// REQUIRES: This version has been saved (see VersionSet::SaveTo) // REQUIRES: This version has been saved (see VersionSet::SaveTo)
@ -72,6 +66,10 @@ class Version {
// Lookup the value for key. If found, store it in *val and // Lookup the value for key. If found, store it in *val and
// return OK. Else return a non-OK status. Fills *stats. // return OK. Else return a non-OK status. Fills *stats.
// REQUIRES: lock is not held // REQUIRES: lock is not held
struct GetStats {
FileMetaData* seek_file;
int seek_file_level;
};
Status Get(const ReadOptions&, const LookupKey& key, std::string* val, Status Get(const ReadOptions&, const LookupKey& key, std::string* val,
GetStats* stats); GetStats* stats);
@ -93,15 +91,16 @@ class Version {
void GetOverlappingInputs( void GetOverlappingInputs(
int level, int level,
const InternalKey* begin, // nullptr means before all keys const InternalKey* begin, // NULL means before all keys
const InternalKey* end, // nullptr means after all keys const InternalKey* end, // NULL means after all keys
std::vector<FileMetaData*>* inputs); std::vector<FileMetaData*>* inputs);
// Returns true iff some file in the specified level overlaps // Returns true iff some file in the specified level overlaps
// some part of [*smallest_user_key,*largest_user_key]. // some part of [*smallest_user_key,*largest_user_key].
// smallest_user_key==nullptr represents a key smaller than all the DB's keys. // smallest_user_key==NULL represents a key smaller than all keys in the DB.
// largest_user_key==nullptr represents a key largest than all the DB's keys. // largest_user_key==NULL represents a key largest than all keys in the DB.
bool OverlapInLevel(int level, const Slice* smallest_user_key, bool OverlapInLevel(int level,
const Slice* smallest_user_key,
const Slice* largest_user_key); const Slice* largest_user_key);
// Return the level at which we should place a new memtable compaction // Return the level at which we should place a new memtable compaction
@ -119,22 +118,6 @@ class Version {
friend class VersionSet; friend class VersionSet;
class LevelFileNumIterator; class LevelFileNumIterator;
explicit Version(VersionSet* vset)
: vset_(vset),
next_(this),
prev_(this),
refs_(0),
file_to_compact_(nullptr),
file_to_compact_level_(-1),
compaction_score_(-1),
compaction_level_(-1) {}
Version(const Version&) = delete;
Version& operator=(const Version&) = delete;
~Version();
Iterator* NewConcatenatingIterator(const ReadOptions&, int level) const; Iterator* NewConcatenatingIterator(const ReadOptions&, int level) const;
// Call func(arg, level, f) for every file that overlaps user_key in // Call func(arg, level, f) for every file that overlaps user_key in
@ -142,13 +125,14 @@ class Version {
// false, makes no more calls. // false, makes no more calls.
// //
// REQUIRES: user portion of internal_key == user_key. // REQUIRES: user portion of internal_key == user_key.
void ForEachOverlapping(Slice user_key, Slice internal_key, void* arg, void ForEachOverlapping(Slice user_key, Slice internal_key,
void* arg,
bool (*func)(void*, int, FileMetaData*)); bool (*func)(void*, int, FileMetaData*));
VersionSet* vset_; // VersionSet to which this Version belongs VersionSet* vset_; // VersionSet to which this Version belongs
Version* next_; // Next version in linked list Version* next_; // Next version in linked list
Version* prev_; // Previous version in linked list Version* prev_; // Previous version in linked list
int refs_; // Number of live refs to this version int refs_; // Number of live refs to this version
// List of files per level // List of files per level
std::vector<FileMetaData*> files_[config::kNumLevels]; std::vector<FileMetaData*> files_[config::kNumLevels];
@ -162,15 +146,28 @@ class Version {
// are initialized by Finalize(). // are initialized by Finalize().
double compaction_score_; double compaction_score_;
int compaction_level_; int compaction_level_;
explicit Version(VersionSet* vset)
: vset_(vset), next_(this), prev_(this), refs_(0),
file_to_compact_(NULL),
file_to_compact_level_(-1),
compaction_score_(-1),
compaction_level_(-1) {
}
~Version();
// No copying allowed
Version(const Version&);
void operator=(const Version&);
}; };
class VersionSet { class VersionSet {
public: public:
VersionSet(const std::string& dbname, const Options* options, VersionSet(const std::string& dbname,
TableCache* table_cache, const InternalKeyComparator*); const Options* options,
VersionSet(const VersionSet&) = delete; TableCache* table_cache,
VersionSet& operator=(const VersionSet&) = delete; const InternalKeyComparator*);
~VersionSet(); ~VersionSet();
// Apply *edit to the current version to form a new descriptor that // Apply *edit to the current version to form a new descriptor that
@ -182,7 +179,7 @@ class VersionSet {
EXCLUSIVE_LOCKS_REQUIRED(mu); EXCLUSIVE_LOCKS_REQUIRED(mu);
// Recover the last saved descriptor from persistent storage. // Recover the last saved descriptor from persistent storage.
Status Recover(bool* save_manifest); Status Recover(bool *save_manifest);
// Return the current version. // Return the current version.
Version* current() const { return current_; } Version* current() const { return current_; }
@ -228,17 +225,19 @@ class VersionSet {
uint64_t PrevLogNumber() const { return prev_log_number_; } uint64_t PrevLogNumber() const { return prev_log_number_; }
// Pick level and inputs for a new compaction. // Pick level and inputs for a new compaction.
// Returns nullptr if there is no compaction to be done. // Returns NULL if there is no compaction to be done.
// Otherwise returns a pointer to a heap-allocated object that // Otherwise returns a pointer to a heap-allocated object that
// describes the compaction. Caller should delete the result. // describes the compaction. Caller should delete the result.
Compaction* PickCompaction(); Compaction* PickCompaction();
// Return a compaction object for compacting the range [begin,end] in // Return a compaction object for compacting the range [begin,end] in
// the specified level. Returns nullptr if there is nothing in that // the specified level. Returns NULL if there is nothing in that
// level that overlaps the specified range. Caller should delete // level that overlaps the specified range. Caller should delete
// the result. // the result.
Compaction* CompactRange(int level, const InternalKey* begin, Compaction* CompactRange(
const InternalKey* end); int level,
const InternalKey* begin,
const InternalKey* end);
// Return the maximum overlapping data (in bytes) at next level for any // Return the maximum overlapping data (in bytes) at next level for any
// file at a level >= 1. // file at a level >= 1.
@ -251,7 +250,7 @@ class VersionSet {
// Returns true iff some level needs a compaction. // Returns true iff some level needs a compaction.
bool NeedsCompaction() const { bool NeedsCompaction() const {
Version* v = current_; Version* v = current_;
return (v->compaction_score_ >= 1) || (v->file_to_compact_ != nullptr); return (v->compaction_score_ >= 1) || (v->file_to_compact_ != NULL);
} }
// Add all files listed in any live version to *live. // Add all files listed in any live version to *live.
@ -279,12 +278,14 @@ class VersionSet {
void Finalize(Version* v); void Finalize(Version* v);
void GetRange(const std::vector<FileMetaData*>& inputs, InternalKey* smallest, void GetRange(const std::vector<FileMetaData*>& inputs,
InternalKey* smallest,
InternalKey* largest); InternalKey* largest);
void GetRange2(const std::vector<FileMetaData*>& inputs1, void GetRange2(const std::vector<FileMetaData*>& inputs1,
const std::vector<FileMetaData*>& inputs2, const std::vector<FileMetaData*>& inputs2,
InternalKey* smallest, InternalKey* largest); InternalKey* smallest,
InternalKey* largest);
void SetupOtherInputs(Compaction* c); void SetupOtherInputs(Compaction* c);
@ -313,6 +314,10 @@ class VersionSet {
// Per-level key at which the next compaction at that level should start. // Per-level key at which the next compaction at that level should start.
// Either an empty string, or a valid InternalKey. // Either an empty string, or a valid InternalKey.
std::string compact_pointer_[config::kNumLevels]; std::string compact_pointer_[config::kNumLevels];
// No copying allowed
VersionSet(const VersionSet&);
void operator=(const VersionSet&);
}; };
// A Compaction encapsulates information about a compaction. // A Compaction encapsulates information about a compaction.
@ -361,7 +366,7 @@ class Compaction {
friend class Version; friend class Version;
friend class VersionSet; friend class VersionSet;
Compaction(const Options* options, int level); explicit Compaction(int level);
int level_; int level_;
uint64_t max_output_file_size_; uint64_t max_output_file_size_;
@ -369,9 +374,9 @@ class Compaction {
VersionEdit edit_; VersionEdit edit_;
// Each compaction reads inputs from "level_" and "level_+1" // Each compaction reads inputs from "level_" and "level_+1"
std::vector<FileMetaData*> inputs_[2]; // The two sets of inputs std::vector<FileMetaData*> inputs_[2]; // The two sets of inputs
// State used to check for number of overlapping grandparent files // State used to check for number of of overlapping grandparent files
// (parent == level_ + 1, grandparent == level_ + 2) // (parent == level_ + 1, grandparent == level_ + 2)
std::vector<FileMetaData*> grandparents_; std::vector<FileMetaData*> grandparents_;
size_t grandparent_index_; // Index in grandparent_starts_ size_t grandparent_index_; // Index in grandparent_starts_

View File

@ -3,16 +3,18 @@
// found in the LICENSE file. See the AUTHORS file for names of contributors. // found in the LICENSE file. See the AUTHORS file for names of contributors.
#include "db/version_set.h" #include "db/version_set.h"
#include "gtest/gtest.h"
#include "util/logging.h" #include "util/logging.h"
#include "util/testharness.h"
#include "util/testutil.h" #include "util/testutil.h"
namespace leveldb { namespace leveldb {
class FindFileTest : public testing::Test { class FindFileTest {
public: public:
FindFileTest() : disjoint_sorted_files_(true) {} std::vector<FileMetaData*> files_;
bool disjoint_sorted_files_;
FindFileTest() : disjoint_sorted_files_(true) { }
~FindFileTest() { ~FindFileTest() {
for (int i = 0; i < files_.size(); i++) { for (int i = 0; i < files_.size(); i++) {
@ -38,28 +40,23 @@ class FindFileTest : public testing::Test {
bool Overlaps(const char* smallest, const char* largest) { bool Overlaps(const char* smallest, const char* largest) {
InternalKeyComparator cmp(BytewiseComparator()); InternalKeyComparator cmp(BytewiseComparator());
Slice s(smallest != nullptr ? smallest : ""); Slice s(smallest != NULL ? smallest : "");
Slice l(largest != nullptr ? largest : ""); Slice l(largest != NULL ? largest : "");
return SomeFileOverlapsRange(cmp, disjoint_sorted_files_, files_, return SomeFileOverlapsRange(cmp, disjoint_sorted_files_, files_,
(smallest != nullptr ? &s : nullptr), (smallest != NULL ? &s : NULL),
(largest != nullptr ? &l : nullptr)); (largest != NULL ? &l : NULL));
} }
bool disjoint_sorted_files_;
private:
std::vector<FileMetaData*> files_;
}; };
TEST_F(FindFileTest, Empty) { TEST(FindFileTest, Empty) {
ASSERT_EQ(0, Find("foo")); ASSERT_EQ(0, Find("foo"));
ASSERT_TRUE(!Overlaps("a", "z")); ASSERT_TRUE(! Overlaps("a", "z"));
ASSERT_TRUE(!Overlaps(nullptr, "z")); ASSERT_TRUE(! Overlaps(NULL, "z"));
ASSERT_TRUE(!Overlaps("a", nullptr)); ASSERT_TRUE(! Overlaps("a", NULL));
ASSERT_TRUE(!Overlaps(nullptr, nullptr)); ASSERT_TRUE(! Overlaps(NULL, NULL));
} }
TEST_F(FindFileTest, Single) { TEST(FindFileTest, Single) {
Add("p", "q"); Add("p", "q");
ASSERT_EQ(0, Find("a")); ASSERT_EQ(0, Find("a"));
ASSERT_EQ(0, Find("p")); ASSERT_EQ(0, Find("p"));
@ -68,8 +65,8 @@ TEST_F(FindFileTest, Single) {
ASSERT_EQ(1, Find("q1")); ASSERT_EQ(1, Find("q1"));
ASSERT_EQ(1, Find("z")); ASSERT_EQ(1, Find("z"));
ASSERT_TRUE(!Overlaps("a", "b")); ASSERT_TRUE(! Overlaps("a", "b"));
ASSERT_TRUE(!Overlaps("z1", "z2")); ASSERT_TRUE(! Overlaps("z1", "z2"));
ASSERT_TRUE(Overlaps("a", "p")); ASSERT_TRUE(Overlaps("a", "p"));
ASSERT_TRUE(Overlaps("a", "q")); ASSERT_TRUE(Overlaps("a", "q"));
ASSERT_TRUE(Overlaps("a", "z")); ASSERT_TRUE(Overlaps("a", "z"));
@ -81,15 +78,16 @@ TEST_F(FindFileTest, Single) {
ASSERT_TRUE(Overlaps("q", "q")); ASSERT_TRUE(Overlaps("q", "q"));
ASSERT_TRUE(Overlaps("q", "q1")); ASSERT_TRUE(Overlaps("q", "q1"));
ASSERT_TRUE(!Overlaps(nullptr, "j")); ASSERT_TRUE(! Overlaps(NULL, "j"));
ASSERT_TRUE(!Overlaps("r", nullptr)); ASSERT_TRUE(! Overlaps("r", NULL));
ASSERT_TRUE(Overlaps(nullptr, "p")); ASSERT_TRUE(Overlaps(NULL, "p"));
ASSERT_TRUE(Overlaps(nullptr, "p1")); ASSERT_TRUE(Overlaps(NULL, "p1"));
ASSERT_TRUE(Overlaps("q", nullptr)); ASSERT_TRUE(Overlaps("q", NULL));
ASSERT_TRUE(Overlaps(nullptr, nullptr)); ASSERT_TRUE(Overlaps(NULL, NULL));
} }
TEST_F(FindFileTest, Multiple) {
TEST(FindFileTest, Multiple) {
Add("150", "200"); Add("150", "200");
Add("200", "250"); Add("200", "250");
Add("300", "350"); Add("300", "350");
@ -112,10 +110,10 @@ TEST_F(FindFileTest, Multiple) {
ASSERT_EQ(3, Find("450")); ASSERT_EQ(3, Find("450"));
ASSERT_EQ(4, Find("451")); ASSERT_EQ(4, Find("451"));
ASSERT_TRUE(!Overlaps("100", "149")); ASSERT_TRUE(! Overlaps("100", "149"));
ASSERT_TRUE(!Overlaps("251", "299")); ASSERT_TRUE(! Overlaps("251", "299"));
ASSERT_TRUE(!Overlaps("451", "500")); ASSERT_TRUE(! Overlaps("451", "500"));
ASSERT_TRUE(!Overlaps("351", "399")); ASSERT_TRUE(! Overlaps("351", "399"));
ASSERT_TRUE(Overlaps("100", "150")); ASSERT_TRUE(Overlaps("100", "150"));
ASSERT_TRUE(Overlaps("100", "200")); ASSERT_TRUE(Overlaps("100", "200"));
@ -127,41 +125,41 @@ TEST_F(FindFileTest, Multiple) {
ASSERT_TRUE(Overlaps("450", "500")); ASSERT_TRUE(Overlaps("450", "500"));
} }
TEST_F(FindFileTest, MultipleNullBoundaries) { TEST(FindFileTest, MultipleNullBoundaries) {
Add("150", "200"); Add("150", "200");
Add("200", "250"); Add("200", "250");
Add("300", "350"); Add("300", "350");
Add("400", "450"); Add("400", "450");
ASSERT_TRUE(!Overlaps(nullptr, "149")); ASSERT_TRUE(! Overlaps(NULL, "149"));
ASSERT_TRUE(!Overlaps("451", nullptr)); ASSERT_TRUE(! Overlaps("451", NULL));
ASSERT_TRUE(Overlaps(nullptr, nullptr)); ASSERT_TRUE(Overlaps(NULL, NULL));
ASSERT_TRUE(Overlaps(nullptr, "150")); ASSERT_TRUE(Overlaps(NULL, "150"));
ASSERT_TRUE(Overlaps(nullptr, "199")); ASSERT_TRUE(Overlaps(NULL, "199"));
ASSERT_TRUE(Overlaps(nullptr, "200")); ASSERT_TRUE(Overlaps(NULL, "200"));
ASSERT_TRUE(Overlaps(nullptr, "201")); ASSERT_TRUE(Overlaps(NULL, "201"));
ASSERT_TRUE(Overlaps(nullptr, "400")); ASSERT_TRUE(Overlaps(NULL, "400"));
ASSERT_TRUE(Overlaps(nullptr, "800")); ASSERT_TRUE(Overlaps(NULL, "800"));
ASSERT_TRUE(Overlaps("100", nullptr)); ASSERT_TRUE(Overlaps("100", NULL));
ASSERT_TRUE(Overlaps("200", nullptr)); ASSERT_TRUE(Overlaps("200", NULL));
ASSERT_TRUE(Overlaps("449", nullptr)); ASSERT_TRUE(Overlaps("449", NULL));
ASSERT_TRUE(Overlaps("450", nullptr)); ASSERT_TRUE(Overlaps("450", NULL));
} }
TEST_F(FindFileTest, OverlapSequenceChecks) { TEST(FindFileTest, OverlapSequenceChecks) {
Add("200", "200", 5000, 3000); Add("200", "200", 5000, 3000);
ASSERT_TRUE(!Overlaps("199", "199")); ASSERT_TRUE(! Overlaps("199", "199"));
ASSERT_TRUE(!Overlaps("201", "300")); ASSERT_TRUE(! Overlaps("201", "300"));
ASSERT_TRUE(Overlaps("200", "200")); ASSERT_TRUE(Overlaps("200", "200"));
ASSERT_TRUE(Overlaps("190", "200")); ASSERT_TRUE(Overlaps("190", "200"));
ASSERT_TRUE(Overlaps("200", "210")); ASSERT_TRUE(Overlaps("200", "210"));
} }
TEST_F(FindFileTest, OverlappingFiles) { TEST(FindFileTest, OverlappingFiles) {
Add("150", "600"); Add("150", "600");
Add("400", "500"); Add("400", "500");
disjoint_sorted_files_ = false; disjoint_sorted_files_ = false;
ASSERT_TRUE(!Overlaps("100", "149")); ASSERT_TRUE(! Overlaps("100", "149"));
ASSERT_TRUE(!Overlaps("601", "700")); ASSERT_TRUE(! Overlaps("601", "700"));
ASSERT_TRUE(Overlaps("100", "150")); ASSERT_TRUE(Overlaps("100", "150"));
ASSERT_TRUE(Overlaps("100", "200")); ASSERT_TRUE(Overlaps("100", "200"));
ASSERT_TRUE(Overlaps("100", "300")); ASSERT_TRUE(Overlaps("100", "300"));
@ -174,158 +172,8 @@ TEST_F(FindFileTest, OverlappingFiles) {
ASSERT_TRUE(Overlaps("600", "700")); ASSERT_TRUE(Overlaps("600", "700"));
} }
void AddBoundaryInputs(const InternalKeyComparator& icmp,
const std::vector<FileMetaData*>& level_files,
std::vector<FileMetaData*>* compaction_files);
class AddBoundaryInputsTest : public testing::Test {
public:
std::vector<FileMetaData*> level_files_;
std::vector<FileMetaData*> compaction_files_;
std::vector<FileMetaData*> all_files_;
InternalKeyComparator icmp_;
AddBoundaryInputsTest() : icmp_(BytewiseComparator()) {}
~AddBoundaryInputsTest() {
for (size_t i = 0; i < all_files_.size(); ++i) {
delete all_files_[i];
}
all_files_.clear();
}
FileMetaData* CreateFileMetaData(uint64_t number, InternalKey smallest,
InternalKey largest) {
FileMetaData* f = new FileMetaData();
f->number = number;
f->smallest = smallest;
f->largest = largest;
all_files_.push_back(f);
return f;
}
};
TEST_F(AddBoundaryInputsTest, TestEmptyFileSets) {
AddBoundaryInputs(icmp_, level_files_, &compaction_files_);
ASSERT_TRUE(compaction_files_.empty());
ASSERT_TRUE(level_files_.empty());
}
TEST_F(AddBoundaryInputsTest, TestEmptyLevelFiles) {
FileMetaData* f1 =
CreateFileMetaData(1, InternalKey("100", 2, kTypeValue),
InternalKey(InternalKey("100", 1, kTypeValue)));
compaction_files_.push_back(f1);
AddBoundaryInputs(icmp_, level_files_, &compaction_files_);
ASSERT_EQ(1, compaction_files_.size());
ASSERT_EQ(f1, compaction_files_[0]);
ASSERT_TRUE(level_files_.empty());
}
TEST_F(AddBoundaryInputsTest, TestEmptyCompactionFiles) {
FileMetaData* f1 =
CreateFileMetaData(1, InternalKey("100", 2, kTypeValue),
InternalKey(InternalKey("100", 1, kTypeValue)));
level_files_.push_back(f1);
AddBoundaryInputs(icmp_, level_files_, &compaction_files_);
ASSERT_TRUE(compaction_files_.empty());
ASSERT_EQ(1, level_files_.size());
ASSERT_EQ(f1, level_files_[0]);
}
TEST_F(AddBoundaryInputsTest, TestNoBoundaryFiles) {
FileMetaData* f1 =
CreateFileMetaData(1, InternalKey("100", 2, kTypeValue),
InternalKey(InternalKey("100", 1, kTypeValue)));
FileMetaData* f2 =
CreateFileMetaData(1, InternalKey("200", 2, kTypeValue),
InternalKey(InternalKey("200", 1, kTypeValue)));
FileMetaData* f3 =
CreateFileMetaData(1, InternalKey("300", 2, kTypeValue),
InternalKey(InternalKey("300", 1, kTypeValue)));
level_files_.push_back(f3);
level_files_.push_back(f2);
level_files_.push_back(f1);
compaction_files_.push_back(f2);
compaction_files_.push_back(f3);
AddBoundaryInputs(icmp_, level_files_, &compaction_files_);
ASSERT_EQ(2, compaction_files_.size());
}
TEST_F(AddBoundaryInputsTest, TestOneBoundaryFiles) {
FileMetaData* f1 =
CreateFileMetaData(1, InternalKey("100", 3, kTypeValue),
InternalKey(InternalKey("100", 2, kTypeValue)));
FileMetaData* f2 =
CreateFileMetaData(1, InternalKey("100", 1, kTypeValue),
InternalKey(InternalKey("200", 3, kTypeValue)));
FileMetaData* f3 =
CreateFileMetaData(1, InternalKey("300", 2, kTypeValue),
InternalKey(InternalKey("300", 1, kTypeValue)));
level_files_.push_back(f3);
level_files_.push_back(f2);
level_files_.push_back(f1);
compaction_files_.push_back(f1);
AddBoundaryInputs(icmp_, level_files_, &compaction_files_);
ASSERT_EQ(2, compaction_files_.size());
ASSERT_EQ(f1, compaction_files_[0]);
ASSERT_EQ(f2, compaction_files_[1]);
}
TEST_F(AddBoundaryInputsTest, TestTwoBoundaryFiles) {
FileMetaData* f1 =
CreateFileMetaData(1, InternalKey("100", 6, kTypeValue),
InternalKey(InternalKey("100", 5, kTypeValue)));
FileMetaData* f2 =
CreateFileMetaData(1, InternalKey("100", 2, kTypeValue),
InternalKey(InternalKey("300", 1, kTypeValue)));
FileMetaData* f3 =
CreateFileMetaData(1, InternalKey("100", 4, kTypeValue),
InternalKey(InternalKey("100", 3, kTypeValue)));
level_files_.push_back(f2);
level_files_.push_back(f3);
level_files_.push_back(f1);
compaction_files_.push_back(f1);
AddBoundaryInputs(icmp_, level_files_, &compaction_files_);
ASSERT_EQ(3, compaction_files_.size());
ASSERT_EQ(f1, compaction_files_[0]);
ASSERT_EQ(f3, compaction_files_[1]);
ASSERT_EQ(f2, compaction_files_[2]);
}
TEST_F(AddBoundaryInputsTest, TestDisjoinFilePointers) {
FileMetaData* f1 =
CreateFileMetaData(1, InternalKey("100", 6, kTypeValue),
InternalKey(InternalKey("100", 5, kTypeValue)));
FileMetaData* f2 =
CreateFileMetaData(1, InternalKey("100", 6, kTypeValue),
InternalKey(InternalKey("100", 5, kTypeValue)));
FileMetaData* f3 =
CreateFileMetaData(1, InternalKey("100", 2, kTypeValue),
InternalKey(InternalKey("300", 1, kTypeValue)));
FileMetaData* f4 =
CreateFileMetaData(1, InternalKey("100", 4, kTypeValue),
InternalKey(InternalKey("100", 3, kTypeValue)));
level_files_.push_back(f2);
level_files_.push_back(f3);
level_files_.push_back(f4);
compaction_files_.push_back(f1);
AddBoundaryInputs(icmp_, level_files_, &compaction_files_);
ASSERT_EQ(3, compaction_files_.size());
ASSERT_EQ(f1, compaction_files_[0]);
ASSERT_EQ(f4, compaction_files_[1]);
ASSERT_EQ(f3, compaction_files_[2]);
}
} // namespace leveldb } // namespace leveldb
int main(int argc, char** argv) {
return leveldb::test::RunAllTests();
}

View File

@ -15,10 +15,10 @@
#include "leveldb/write_batch.h" #include "leveldb/write_batch.h"
#include "leveldb/db.h"
#include "db/dbformat.h" #include "db/dbformat.h"
#include "db/memtable.h" #include "db/memtable.h"
#include "db/write_batch_internal.h" #include "db/write_batch_internal.h"
#include "leveldb/db.h"
#include "util/coding.h" #include "util/coding.h"
namespace leveldb { namespace leveldb {
@ -26,19 +26,19 @@ namespace leveldb {
// WriteBatch header has an 8-byte sequence number followed by a 4-byte count. // WriteBatch header has an 8-byte sequence number followed by a 4-byte count.
static const size_t kHeader = 12; static const size_t kHeader = 12;
WriteBatch::WriteBatch() { Clear(); } WriteBatch::WriteBatch() {
Clear();
}
WriteBatch::~WriteBatch() = default; WriteBatch::~WriteBatch() { }
WriteBatch::Handler::~Handler() = default; WriteBatch::Handler::~Handler() { }
void WriteBatch::Clear() { void WriteBatch::Clear() {
rep_.clear(); rep_.clear();
rep_.resize(kHeader); rep_.resize(kHeader);
} }
size_t WriteBatch::ApproximateSize() const { return rep_.size(); }
Status WriteBatch::Iterate(Handler* handler) const { Status WriteBatch::Iterate(Handler* handler) const {
Slice input(rep_); Slice input(rep_);
if (input.size() < kHeader) { if (input.size() < kHeader) {
@ -108,28 +108,25 @@ void WriteBatch::Delete(const Slice& key) {
PutLengthPrefixedSlice(&rep_, key); PutLengthPrefixedSlice(&rep_, key);
} }
void WriteBatch::Append(const WriteBatch& source) {
WriteBatchInternal::Append(this, &source);
}
namespace { namespace {
class MemTableInserter : public WriteBatch::Handler { class MemTableInserter : public WriteBatch::Handler {
public: public:
SequenceNumber sequence_; SequenceNumber sequence_;
MemTable* mem_; MemTable* mem_;
void Put(const Slice& key, const Slice& value) override { virtual void Put(const Slice& key, const Slice& value) {
mem_->Add(sequence_, kTypeValue, key, value); mem_->Add(sequence_, kTypeValue, key, value);
sequence_++; sequence_++;
} }
void Delete(const Slice& key) override { virtual void Delete(const Slice& key) {
mem_->Add(sequence_, kTypeDeletion, key, Slice()); mem_->Add(sequence_, kTypeDeletion, key, Slice());
sequence_++; sequence_++;
} }
}; };
} // namespace } // namespace
Status WriteBatchInternal::InsertInto(const WriteBatch* b, MemTable* memtable) { Status WriteBatchInternal::InsertInto(const WriteBatch* b,
MemTable* memtable) {
MemTableInserter inserter; MemTableInserter inserter;
inserter.sequence_ = WriteBatchInternal::Sequence(b); inserter.sequence_ = WriteBatchInternal::Sequence(b);
inserter.mem_ = memtable; inserter.mem_ = memtable;

View File

@ -29,9 +29,13 @@ class WriteBatchInternal {
// this batch. // this batch.
static void SetSequence(WriteBatch* batch, SequenceNumber seq); static void SetSequence(WriteBatch* batch, SequenceNumber seq);
static Slice Contents(const WriteBatch* batch) { return Slice(batch->rep_); } static Slice Contents(const WriteBatch* batch) {
return Slice(batch->rep_);
}
static size_t ByteSize(const WriteBatch* batch) { return batch->rep_.size(); } static size_t ByteSize(const WriteBatch* batch) {
return batch->rep_.size();
}
static void SetContents(WriteBatch* batch, const Slice& contents); static void SetContents(WriteBatch* batch, const Slice& contents);
@ -42,4 +46,5 @@ class WriteBatchInternal {
} // namespace leveldb } // namespace leveldb
#endif // STORAGE_LEVELDB_DB_WRITE_BATCH_INTERNAL_H_ #endif // STORAGE_LEVELDB_DB_WRITE_BATCH_INTERNAL_H_

View File

@ -2,12 +2,13 @@
// Use of this source code is governed by a BSD-style license that can be // Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file. See the AUTHORS file for names of contributors. // found in the LICENSE file. See the AUTHORS file for names of contributors.
#include "gtest/gtest.h" #include "leveldb/db.h"
#include "db/memtable.h" #include "db/memtable.h"
#include "db/write_batch_internal.h" #include "db/write_batch_internal.h"
#include "leveldb/db.h"
#include "leveldb/env.h" #include "leveldb/env.h"
#include "util/logging.h" #include "util/logging.h"
#include "util/testharness.h"
namespace leveldb { namespace leveldb {
@ -21,7 +22,7 @@ static std::string PrintContents(WriteBatch* b) {
Iterator* iter = mem->NewIterator(); Iterator* iter = mem->NewIterator();
for (iter->SeekToFirst(); iter->Valid(); iter->Next()) { for (iter->SeekToFirst(); iter->Valid(); iter->Next()) {
ParsedInternalKey ikey; ParsedInternalKey ikey;
EXPECT_TRUE(ParseInternalKey(iter->key(), &ikey)); ASSERT_TRUE(ParseInternalKey(iter->key(), &ikey));
switch (ikey.type) { switch (ikey.type) {
case kTypeValue: case kTypeValue:
state.append("Put("); state.append("Put(");
@ -51,6 +52,8 @@ static std::string PrintContents(WriteBatch* b) {
return state; return state;
} }
class WriteBatchTest { };
TEST(WriteBatchTest, Empty) { TEST(WriteBatchTest, Empty) {
WriteBatch batch; WriteBatch batch;
ASSERT_EQ("", PrintContents(&batch)); ASSERT_EQ("", PrintContents(&batch));
@ -65,11 +68,10 @@ TEST(WriteBatchTest, Multiple) {
WriteBatchInternal::SetSequence(&batch, 100); WriteBatchInternal::SetSequence(&batch, 100);
ASSERT_EQ(100, WriteBatchInternal::Sequence(&batch)); ASSERT_EQ(100, WriteBatchInternal::Sequence(&batch));
ASSERT_EQ(3, WriteBatchInternal::Count(&batch)); ASSERT_EQ(3, WriteBatchInternal::Count(&batch));
ASSERT_EQ( ASSERT_EQ("Put(baz, boo)@102"
"Put(baz, boo)@102" "Delete(box)@101"
"Delete(box)@101" "Put(foo, bar)@100",
"Put(foo, bar)@100", PrintContents(&batch));
PrintContents(&batch));
} }
TEST(WriteBatchTest, Corruption) { TEST(WriteBatchTest, Corruption) {
@ -79,54 +81,40 @@ TEST(WriteBatchTest, Corruption) {
WriteBatchInternal::SetSequence(&batch, 200); WriteBatchInternal::SetSequence(&batch, 200);
Slice contents = WriteBatchInternal::Contents(&batch); Slice contents = WriteBatchInternal::Contents(&batch);
WriteBatchInternal::SetContents(&batch, WriteBatchInternal::SetContents(&batch,
Slice(contents.data(), contents.size() - 1)); Slice(contents.data(),contents.size()-1));
ASSERT_EQ( ASSERT_EQ("Put(foo, bar)@200"
"Put(foo, bar)@200" "ParseError()",
"ParseError()", PrintContents(&batch));
PrintContents(&batch));
} }
TEST(WriteBatchTest, Append) { TEST(WriteBatchTest, Append) {
WriteBatch b1, b2; WriteBatch b1, b2;
WriteBatchInternal::SetSequence(&b1, 200); WriteBatchInternal::SetSequence(&b1, 200);
WriteBatchInternal::SetSequence(&b2, 300); WriteBatchInternal::SetSequence(&b2, 300);
b1.Append(b2); WriteBatchInternal::Append(&b1, &b2);
ASSERT_EQ("", PrintContents(&b1)); ASSERT_EQ("",
PrintContents(&b1));
b2.Put("a", "va"); b2.Put("a", "va");
b1.Append(b2); WriteBatchInternal::Append(&b1, &b2);
ASSERT_EQ("Put(a, va)@200", PrintContents(&b1)); ASSERT_EQ("Put(a, va)@200",
PrintContents(&b1));
b2.Clear(); b2.Clear();
b2.Put("b", "vb"); b2.Put("b", "vb");
b1.Append(b2); WriteBatchInternal::Append(&b1, &b2);
ASSERT_EQ( ASSERT_EQ("Put(a, va)@200"
"Put(a, va)@200" "Put(b, vb)@201",
"Put(b, vb)@201", PrintContents(&b1));
PrintContents(&b1));
b2.Delete("foo"); b2.Delete("foo");
b1.Append(b2); WriteBatchInternal::Append(&b1, &b2);
ASSERT_EQ( ASSERT_EQ("Put(a, va)@200"
"Put(a, va)@200" "Put(b, vb)@202"
"Put(b, vb)@202" "Put(b, vb)@201"
"Put(b, vb)@201" "Delete(foo)@203",
"Delete(foo)@203", PrintContents(&b1));
PrintContents(&b1));
}
TEST(WriteBatchTest, ApproximateSize) {
WriteBatch batch;
size_t empty_size = batch.ApproximateSize();
batch.Put(Slice("foo"), Slice("bar"));
size_t one_key_size = batch.ApproximateSize();
ASSERT_LT(empty_size, one_key_size);
batch.Put(Slice("baz"), Slice("boo"));
size_t two_keys_size = batch.ApproximateSize();
ASSERT_LT(one_key_size, two_keys_size);
batch.Delete(Slice("box"));
size_t post_delete_size = batch.ApproximateSize();
ASSERT_LT(two_keys_size, post_delete_size);
} }
} // namespace leveldb } // namespace leveldb
int main(int argc, char** argv) {
return leveldb::test::RunAllTests();
}

View File

@ -2,11 +2,9 @@
// Use of this source code is governed by a BSD-style license that can be // Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file. See the AUTHORS file for names of contributors. // found in the LICENSE file. See the AUTHORS file for names of contributors.
#include <stdio.h>
#include <stdlib.h>
#include <sqlite3.h> #include <sqlite3.h>
#include <cstdio>
#include <cstdlib>
#include "util/histogram.h" #include "util/histogram.h"
#include "util/random.h" #include "util/random.h"
#include "util/testutil.h" #include "util/testutil.h"
@ -40,7 +38,8 @@ static const char* FLAGS_benchmarks =
"fillrand100K," "fillrand100K,"
"fillseq100K," "fillseq100K,"
"readseq," "readseq,"
"readrand100K,"; "readrand100K,"
;
// Number of key/values to place in database // Number of key/values to place in database
static int FLAGS_num = 1000000; static int FLAGS_num = 1000000;
@ -70,9 +69,6 @@ static int FLAGS_num_pages = 4096;
// benchmark will fail. // benchmark will fail.
static bool FLAGS_use_existing_db = false; static bool FLAGS_use_existing_db = false;
// If true, the SQLite table has ROWIDs.
static bool FLAGS_use_rowids = false;
// If true, we allow batch writes to occur // If true, we allow batch writes to occur
static bool FLAGS_transaction = true; static bool FLAGS_transaction = true;
@ -80,35 +76,38 @@ static bool FLAGS_transaction = true;
static bool FLAGS_WAL_enabled = true; static bool FLAGS_WAL_enabled = true;
// Use the db with the following name. // Use the db with the following name.
static const char* FLAGS_db = nullptr; static const char* FLAGS_db = NULL;
inline static void ExecErrorCheck(int status, char* err_msg) { inline
static void ExecErrorCheck(int status, char *err_msg) {
if (status != SQLITE_OK) { if (status != SQLITE_OK) {
std::fprintf(stderr, "SQL error: %s\n", err_msg); fprintf(stderr, "SQL error: %s\n", err_msg);
sqlite3_free(err_msg); sqlite3_free(err_msg);
std::exit(1); exit(1);
} }
} }
inline static void StepErrorCheck(int status) { inline
static void StepErrorCheck(int status) {
if (status != SQLITE_DONE) { if (status != SQLITE_DONE) {
std::fprintf(stderr, "SQL step error: status = %d\n", status); fprintf(stderr, "SQL step error: status = %d\n", status);
std::exit(1); exit(1);
} }
} }
inline static void ErrorCheck(int status) { inline
static void ErrorCheck(int status) {
if (status != SQLITE_OK) { if (status != SQLITE_OK) {
std::fprintf(stderr, "sqlite3 error: status = %d\n", status); fprintf(stderr, "sqlite3 error: status = %d\n", status);
std::exit(1); exit(1);
} }
} }
inline static void WalCheckpoint(sqlite3* db_) { inline
static void WalCheckpoint(sqlite3* db_) {
// Flush all writes to disk // Flush all writes to disk
if (FLAGS_WAL_enabled) { if (FLAGS_WAL_enabled) {
sqlite3_wal_checkpoint_v2(db_, nullptr, SQLITE_CHECKPOINT_FULL, nullptr, sqlite3_wal_checkpoint_v2(db_, NULL, SQLITE_CHECKPOINT_FULL, NULL, NULL);
nullptr);
} }
} }
@ -153,7 +152,7 @@ static Slice TrimSpace(Slice s) {
start++; start++;
} }
int limit = s.size(); int limit = s.size();
while (limit > start && isspace(s[limit - 1])) { while (limit > start && isspace(s[limit-1])) {
limit--; limit--;
} }
return Slice(s.data() + start, limit - start); return Slice(s.data() + start, limit - start);
@ -177,51 +176,49 @@ class Benchmark {
// State kept for progress messages // State kept for progress messages
int done_; int done_;
int next_report_; // When to report next int next_report_; // When to report next
void PrintHeader() { void PrintHeader() {
const int kKeySize = 16; const int kKeySize = 16;
PrintEnvironment(); PrintEnvironment();
std::fprintf(stdout, "Keys: %d bytes each\n", kKeySize); fprintf(stdout, "Keys: %d bytes each\n", kKeySize);
std::fprintf(stdout, "Values: %d bytes each\n", FLAGS_value_size); fprintf(stdout, "Values: %d bytes each\n", FLAGS_value_size);
std::fprintf(stdout, "Entries: %d\n", num_); fprintf(stdout, "Entries: %d\n", num_);
std::fprintf(stdout, "RawSize: %.1f MB (estimated)\n", fprintf(stdout, "RawSize: %.1f MB (estimated)\n",
((static_cast<int64_t>(kKeySize + FLAGS_value_size) * num_) / ((static_cast<int64_t>(kKeySize + FLAGS_value_size) * num_)
1048576.0)); / 1048576.0));
PrintWarnings(); PrintWarnings();
std::fprintf(stdout, "------------------------------------------------\n"); fprintf(stdout, "------------------------------------------------\n");
} }
void PrintWarnings() { void PrintWarnings() {
#if defined(__GNUC__) && !defined(__OPTIMIZE__) #if defined(__GNUC__) && !defined(__OPTIMIZE__)
std::fprintf( fprintf(stdout,
stdout, "WARNING: Optimization is disabled: benchmarks unnecessarily slow\n"
"WARNING: Optimization is disabled: benchmarks unnecessarily slow\n"); );
#endif #endif
#ifndef NDEBUG #ifndef NDEBUG
std::fprintf( fprintf(stdout,
stdout, "WARNING: Assertions are enabled; benchmarks unnecessarily slow\n");
"WARNING: Assertions are enabled; benchmarks unnecessarily slow\n");
#endif #endif
} }
void PrintEnvironment() { void PrintEnvironment() {
std::fprintf(stderr, "SQLite: version %s\n", SQLITE_VERSION); fprintf(stderr, "SQLite: version %s\n", SQLITE_VERSION);
#if defined(__linux) #if defined(__linux)
time_t now = time(nullptr); time_t now = time(NULL);
std::fprintf(stderr, "Date: %s", fprintf(stderr, "Date: %s", ctime(&now)); // ctime() adds newline
ctime(&now)); // ctime() adds newline
FILE* cpuinfo = std::fopen("/proc/cpuinfo", "r"); FILE* cpuinfo = fopen("/proc/cpuinfo", "r");
if (cpuinfo != nullptr) { if (cpuinfo != NULL) {
char line[1000]; char line[1000];
int num_cpus = 0; int num_cpus = 0;
std::string cpu_type; std::string cpu_type;
std::string cache_size; std::string cache_size;
while (fgets(line, sizeof(line), cpuinfo) != nullptr) { while (fgets(line, sizeof(line), cpuinfo) != NULL) {
const char* sep = strchr(line, ':'); const char* sep = strchr(line, ':');
if (sep == nullptr) { if (sep == NULL) {
continue; continue;
} }
Slice key = TrimSpace(Slice(line, sep - 1 - line)); Slice key = TrimSpace(Slice(line, sep - 1 - line));
@ -233,9 +230,9 @@ class Benchmark {
cache_size = val.ToString(); cache_size = val.ToString();
} }
} }
std::fclose(cpuinfo); fclose(cpuinfo);
std::fprintf(stderr, "CPU: %d * %s\n", num_cpus, cpu_type.c_str()); fprintf(stderr, "CPU: %d * %s\n", num_cpus, cpu_type.c_str());
std::fprintf(stderr, "CPUCache: %s\n", cache_size.c_str()); fprintf(stderr, "CPUCache: %s\n", cache_size.c_str());
} }
#endif #endif
} }
@ -256,30 +253,23 @@ class Benchmark {
double micros = (now - last_op_finish_) * 1e6; double micros = (now - last_op_finish_) * 1e6;
hist_.Add(micros); hist_.Add(micros);
if (micros > 20000) { if (micros > 20000) {
std::fprintf(stderr, "long op: %.1f micros%30s\r", micros, ""); fprintf(stderr, "long op: %.1f micros%30s\r", micros, "");
std::fflush(stderr); fflush(stderr);
} }
last_op_finish_ = now; last_op_finish_ = now;
} }
done_++; done_++;
if (done_ >= next_report_) { if (done_ >= next_report_) {
if (next_report_ < 1000) if (next_report_ < 1000) next_report_ += 100;
next_report_ += 100; else if (next_report_ < 5000) next_report_ += 500;
else if (next_report_ < 5000) else if (next_report_ < 10000) next_report_ += 1000;
next_report_ += 500; else if (next_report_ < 50000) next_report_ += 5000;
else if (next_report_ < 10000) else if (next_report_ < 100000) next_report_ += 10000;
next_report_ += 1000; else if (next_report_ < 500000) next_report_ += 50000;
else if (next_report_ < 50000) else next_report_ += 100000;
next_report_ += 5000; fprintf(stderr, "... finished %d ops%30s\r", done_, "");
else if (next_report_ < 100000) fflush(stderr);
next_report_ += 10000;
else if (next_report_ < 500000)
next_report_ += 50000;
else
next_report_ += 100000;
std::fprintf(stderr, "... finished %d ops%30s\r", done_, "");
std::fflush(stderr);
} }
} }
@ -292,36 +282,43 @@ class Benchmark {
if (bytes_ > 0) { if (bytes_ > 0) {
char rate[100]; char rate[100];
std::snprintf(rate, sizeof(rate), "%6.1f MB/s", snprintf(rate, sizeof(rate), "%6.1f MB/s",
(bytes_ / 1048576.0) / (finish - start_)); (bytes_ / 1048576.0) / (finish - start_));
if (!message_.empty()) { if (!message_.empty()) {
message_ = std::string(rate) + " " + message_; message_ = std::string(rate) + " " + message_;
} else { } else {
message_ = rate; message_ = rate;
} }
} }
std::fprintf(stdout, "%-12s : %11.3f micros/op;%s%s\n", fprintf(stdout, "%-12s : %11.3f micros/op;%s%s\n",
name.ToString().c_str(), (finish - start_) * 1e6 / done_, name.ToString().c_str(),
(message_.empty() ? "" : " "), message_.c_str()); (finish - start_) * 1e6 / done_,
(message_.empty() ? "" : " "),
message_.c_str());
if (FLAGS_histogram) { if (FLAGS_histogram) {
std::fprintf(stdout, "Microseconds per op:\n%s\n", fprintf(stdout, "Microseconds per op:\n%s\n", hist_.ToString().c_str());
hist_.ToString().c_str());
} }
std::fflush(stdout); fflush(stdout);
} }
public: public:
enum Order { SEQUENTIAL, RANDOM }; enum Order {
enum DBState { FRESH, EXISTING }; SEQUENTIAL,
RANDOM
};
enum DBState {
FRESH,
EXISTING
};
Benchmark() Benchmark()
: db_(nullptr), : db_(NULL),
db_num_(0), db_num_(0),
num_(FLAGS_num), num_(FLAGS_num),
reads_(FLAGS_reads < 0 ? FLAGS_num : FLAGS_reads), reads_(FLAGS_reads < 0 ? FLAGS_num : FLAGS_reads),
bytes_(0), bytes_(0),
rand_(301) { rand_(301) {
std::vector<std::string> files; std::vector<std::string> files;
std::string test_dir; std::string test_dir;
Env::Default()->GetTestDirectory(&test_dir); Env::Default()->GetTestDirectory(&test_dir);
@ -332,7 +329,7 @@ class Benchmark {
std::string file_name(test_dir); std::string file_name(test_dir);
file_name += "/"; file_name += "/";
file_name += files[i]; file_name += files[i];
Env::Default()->RemoveFile(file_name.c_str()); Env::Default()->DeleteFile(file_name.c_str());
} }
} }
} }
@ -348,12 +345,12 @@ class Benchmark {
Open(); Open();
const char* benchmarks = FLAGS_benchmarks; const char* benchmarks = FLAGS_benchmarks;
while (benchmarks != nullptr) { while (benchmarks != NULL) {
const char* sep = strchr(benchmarks, ','); const char* sep = strchr(benchmarks, ',');
Slice name; Slice name;
if (sep == nullptr) { if (sep == NULL) {
name = benchmarks; name = benchmarks;
benchmarks = nullptr; benchmarks = NULL;
} else { } else {
name = Slice(benchmarks, sep - benchmarks); name = Slice(benchmarks, sep - benchmarks);
benchmarks = sep + 1; benchmarks = sep + 1;
@ -408,8 +405,7 @@ class Benchmark {
} else { } else {
known = false; known = false;
if (name != Slice()) { // No error message for empty name if (name != Slice()) { // No error message for empty name
std::fprintf(stderr, "unknown benchmark '%s'\n", fprintf(stderr, "unknown benchmark '%s'\n", name.ToString().c_str());
name.ToString().c_str());
} }
} }
if (known) { if (known) {
@ -419,37 +415,39 @@ class Benchmark {
} }
void Open() { void Open() {
assert(db_ == nullptr); assert(db_ == NULL);
int status; int status;
char file_name[100]; char file_name[100];
char* err_msg = nullptr; char* err_msg = NULL;
db_num_++; db_num_++;
// Open database // Open database
std::string tmp_dir; std::string tmp_dir;
Env::Default()->GetTestDirectory(&tmp_dir); Env::Default()->GetTestDirectory(&tmp_dir);
std::snprintf(file_name, sizeof(file_name), "%s/dbbench_sqlite3-%d.db", snprintf(file_name, sizeof(file_name),
tmp_dir.c_str(), db_num_); "%s/dbbench_sqlite3-%d.db",
tmp_dir.c_str(),
db_num_);
status = sqlite3_open(file_name, &db_); status = sqlite3_open(file_name, &db_);
if (status) { if (status) {
std::fprintf(stderr, "open error: %s\n", sqlite3_errmsg(db_)); fprintf(stderr, "open error: %s\n", sqlite3_errmsg(db_));
std::exit(1); exit(1);
} }
// Change SQLite cache size // Change SQLite cache size
char cache_size[100]; char cache_size[100];
std::snprintf(cache_size, sizeof(cache_size), "PRAGMA cache_size = %d", snprintf(cache_size, sizeof(cache_size), "PRAGMA cache_size = %d",
FLAGS_num_pages); FLAGS_num_pages);
status = sqlite3_exec(db_, cache_size, nullptr, nullptr, &err_msg); status = sqlite3_exec(db_, cache_size, NULL, NULL, &err_msg);
ExecErrorCheck(status, err_msg); ExecErrorCheck(status, err_msg);
// FLAGS_page_size is defaulted to 1024 // FLAGS_page_size is defaulted to 1024
if (FLAGS_page_size != 1024) { if (FLAGS_page_size != 1024) {
char page_size[100]; char page_size[100];
std::snprintf(page_size, sizeof(page_size), "PRAGMA page_size = %d", snprintf(page_size, sizeof(page_size), "PRAGMA page_size = %d",
FLAGS_page_size); FLAGS_page_size);
status = sqlite3_exec(db_, page_size, nullptr, nullptr, &err_msg); status = sqlite3_exec(db_, page_size, NULL, NULL, &err_msg);
ExecErrorCheck(status, err_msg); ExecErrorCheck(status, err_msg);
} }
@ -459,29 +457,26 @@ class Benchmark {
// LevelDB's default cache size is a combined 4 MB // LevelDB's default cache size is a combined 4 MB
std::string WAL_checkpoint = "PRAGMA wal_autocheckpoint = 4096"; std::string WAL_checkpoint = "PRAGMA wal_autocheckpoint = 4096";
status = sqlite3_exec(db_, WAL_stmt.c_str(), nullptr, nullptr, &err_msg); status = sqlite3_exec(db_, WAL_stmt.c_str(), NULL, NULL, &err_msg);
ExecErrorCheck(status, err_msg); ExecErrorCheck(status, err_msg);
status = status = sqlite3_exec(db_, WAL_checkpoint.c_str(), NULL, NULL, &err_msg);
sqlite3_exec(db_, WAL_checkpoint.c_str(), nullptr, nullptr, &err_msg);
ExecErrorCheck(status, err_msg); ExecErrorCheck(status, err_msg);
} }
// Change locking mode to exclusive and create tables/index for database // Change locking mode to exclusive and create tables/index for database
std::string locking_stmt = "PRAGMA locking_mode = EXCLUSIVE"; std::string locking_stmt = "PRAGMA locking_mode = EXCLUSIVE";
std::string create_stmt = std::string create_stmt =
"CREATE TABLE test (key blob, value blob, PRIMARY KEY(key))"; "CREATE TABLE test (key blob, value blob, PRIMARY KEY(key))";
if (!FLAGS_use_rowids) create_stmt += " WITHOUT ROWID"; std::string stmt_array[] = { locking_stmt, create_stmt };
std::string stmt_array[] = {locking_stmt, create_stmt};
int stmt_array_length = sizeof(stmt_array) / sizeof(std::string); int stmt_array_length = sizeof(stmt_array) / sizeof(std::string);
for (int i = 0; i < stmt_array_length; i++) { for (int i = 0; i < stmt_array_length; i++) {
status = status = sqlite3_exec(db_, stmt_array[i].c_str(), NULL, NULL, &err_msg);
sqlite3_exec(db_, stmt_array[i].c_str(), nullptr, nullptr, &err_msg);
ExecErrorCheck(status, err_msg); ExecErrorCheck(status, err_msg);
} }
} }
void Write(bool write_sync, Order order, DBState state, int num_entries, void Write(bool write_sync, Order order, DBState state,
int value_size, int entries_per_batch) { int num_entries, int value_size, int entries_per_batch) {
// Create new database if state == FRESH // Create new database if state == FRESH
if (state == FRESH) { if (state == FRESH) {
if (FLAGS_use_existing_db) { if (FLAGS_use_existing_db) {
@ -489,18 +484,18 @@ class Benchmark {
return; return;
} }
sqlite3_close(db_); sqlite3_close(db_);
db_ = nullptr; db_ = NULL;
Open(); Open();
Start(); Start();
} }
if (num_entries != num_) { if (num_entries != num_) {
char msg[100]; char msg[100];
std::snprintf(msg, sizeof(msg), "(%d ops)", num_entries); snprintf(msg, sizeof(msg), "(%d ops)", num_entries);
message_ = msg; message_ = msg;
} }
char* err_msg = nullptr; char* err_msg = NULL;
int status; int status;
sqlite3_stmt *replace_stmt, *begin_trans_stmt, *end_trans_stmt; sqlite3_stmt *replace_stmt, *begin_trans_stmt, *end_trans_stmt;
@ -509,20 +504,20 @@ class Benchmark {
std::string end_trans_str = "END TRANSACTION;"; std::string end_trans_str = "END TRANSACTION;";
// Check for synchronous flag in options // Check for synchronous flag in options
std::string sync_stmt = std::string sync_stmt = (write_sync) ? "PRAGMA synchronous = FULL" :
(write_sync) ? "PRAGMA synchronous = FULL" : "PRAGMA synchronous = OFF"; "PRAGMA synchronous = OFF";
status = sqlite3_exec(db_, sync_stmt.c_str(), nullptr, nullptr, &err_msg); status = sqlite3_exec(db_, sync_stmt.c_str(), NULL, NULL, &err_msg);
ExecErrorCheck(status, err_msg); ExecErrorCheck(status, err_msg);
// Preparing sqlite3 statements // Preparing sqlite3 statements
status = sqlite3_prepare_v2(db_, replace_str.c_str(), -1, &replace_stmt, status = sqlite3_prepare_v2(db_, replace_str.c_str(), -1,
nullptr); &replace_stmt, NULL);
ErrorCheck(status); ErrorCheck(status);
status = sqlite3_prepare_v2(db_, begin_trans_str.c_str(), -1, status = sqlite3_prepare_v2(db_, begin_trans_str.c_str(), -1,
&begin_trans_stmt, nullptr); &begin_trans_stmt, NULL);
ErrorCheck(status); ErrorCheck(status);
status = sqlite3_prepare_v2(db_, end_trans_str.c_str(), -1, &end_trans_stmt, status = sqlite3_prepare_v2(db_, end_trans_str.c_str(), -1,
nullptr); &end_trans_stmt, NULL);
ErrorCheck(status); ErrorCheck(status);
bool transaction = (entries_per_batch > 1); bool transaction = (entries_per_batch > 1);
@ -540,16 +535,16 @@ class Benchmark {
const char* value = gen_.Generate(value_size).data(); const char* value = gen_.Generate(value_size).data();
// Create values for key-value pair // Create values for key-value pair
const int k = const int k = (order == SEQUENTIAL) ? i + j :
(order == SEQUENTIAL) ? i + j : (rand_.Next() % num_entries); (rand_.Next() % num_entries);
char key[100]; char key[100];
std::snprintf(key, sizeof(key), "%016d", k); snprintf(key, sizeof(key), "%016d", k);
// Bind KV values into replace_stmt // Bind KV values into replace_stmt
status = sqlite3_bind_blob(replace_stmt, 1, key, 16, SQLITE_STATIC); status = sqlite3_bind_blob(replace_stmt, 1, key, 16, SQLITE_STATIC);
ErrorCheck(status); ErrorCheck(status);
status = sqlite3_bind_blob(replace_stmt, 2, value, value_size, status = sqlite3_bind_blob(replace_stmt, 2, value,
SQLITE_STATIC); value_size, SQLITE_STATIC);
ErrorCheck(status); ErrorCheck(status);
// Execute replace_stmt // Execute replace_stmt
@ -593,12 +588,12 @@ class Benchmark {
// Preparing sqlite3 statements // Preparing sqlite3 statements
status = sqlite3_prepare_v2(db_, begin_trans_str.c_str(), -1, status = sqlite3_prepare_v2(db_, begin_trans_str.c_str(), -1,
&begin_trans_stmt, nullptr); &begin_trans_stmt, NULL);
ErrorCheck(status); ErrorCheck(status);
status = sqlite3_prepare_v2(db_, end_trans_str.c_str(), -1, &end_trans_stmt, status = sqlite3_prepare_v2(db_, end_trans_str.c_str(), -1,
nullptr); &end_trans_stmt, NULL);
ErrorCheck(status); ErrorCheck(status);
status = sqlite3_prepare_v2(db_, read_str.c_str(), -1, &read_stmt, nullptr); status = sqlite3_prepare_v2(db_, read_str.c_str(), -1, &read_stmt, NULL);
ErrorCheck(status); ErrorCheck(status);
bool transaction = (entries_per_batch > 1); bool transaction = (entries_per_batch > 1);
@ -616,15 +611,14 @@ class Benchmark {
// Create key value // Create key value
char key[100]; char key[100];
int k = (order == SEQUENTIAL) ? i + j : (rand_.Next() % reads_); int k = (order == SEQUENTIAL) ? i + j : (rand_.Next() % reads_);
std::snprintf(key, sizeof(key), "%016d", k); snprintf(key, sizeof(key), "%016d", k);
// Bind key value into read_stmt // Bind key value into read_stmt
status = sqlite3_bind_blob(read_stmt, 1, key, 16, SQLITE_STATIC); status = sqlite3_bind_blob(read_stmt, 1, key, 16, SQLITE_STATIC);
ErrorCheck(status); ErrorCheck(status);
// Execute read statement // Execute read statement
while ((status = sqlite3_step(read_stmt)) == SQLITE_ROW) { while ((status = sqlite3_step(read_stmt)) == SQLITE_ROW) {}
}
StepErrorCheck(status); StepErrorCheck(status);
// Reset SQLite statement for another use // Reset SQLite statement for another use
@ -654,10 +648,10 @@ class Benchmark {
void ReadSequential() { void ReadSequential() {
int status; int status;
sqlite3_stmt* pStmt; sqlite3_stmt *pStmt;
std::string read_str = "SELECT * FROM test ORDER BY key"; std::string read_str = "SELECT * FROM test ORDER BY key";
status = sqlite3_prepare_v2(db_, read_str.c_str(), -1, &pStmt, nullptr); status = sqlite3_prepare_v2(db_, read_str.c_str(), -1, &pStmt, NULL);
ErrorCheck(status); ErrorCheck(status);
for (int i = 0; i < reads_ && SQLITE_ROW == sqlite3_step(pStmt); i++) { for (int i = 0; i < reads_ && SQLITE_ROW == sqlite3_step(pStmt); i++) {
bytes_ += sqlite3_column_bytes(pStmt, 1) + sqlite3_column_bytes(pStmt, 2); bytes_ += sqlite3_column_bytes(pStmt, 1) + sqlite3_column_bytes(pStmt, 2);
@ -667,6 +661,7 @@ class Benchmark {
status = sqlite3_finalize(pStmt); status = sqlite3_finalize(pStmt);
ErrorCheck(status); ErrorCheck(status);
} }
}; };
} // namespace leveldb } // namespace leveldb
@ -687,9 +682,6 @@ int main(int argc, char** argv) {
} else if (sscanf(argv[i], "--use_existing_db=%d%c", &n, &junk) == 1 && } else if (sscanf(argv[i], "--use_existing_db=%d%c", &n, &junk) == 1 &&
(n == 0 || n == 1)) { (n == 0 || n == 1)) {
FLAGS_use_existing_db = n; FLAGS_use_existing_db = n;
} else if (sscanf(argv[i], "--use_rowids=%d%c", &n, &junk) == 1 &&
(n == 0 || n == 1)) {
FLAGS_use_rowids = n;
} else if (sscanf(argv[i], "--num=%d%c", &n, &junk) == 1) { } else if (sscanf(argv[i], "--num=%d%c", &n, &junk) == 1) {
FLAGS_num = n; FLAGS_num = n;
} else if (sscanf(argv[i], "--reads=%d%c", &n, &junk) == 1) { } else if (sscanf(argv[i], "--reads=%d%c", &n, &junk) == 1) {
@ -708,16 +700,16 @@ int main(int argc, char** argv) {
} else if (strncmp(argv[i], "--db=", 5) == 0) { } else if (strncmp(argv[i], "--db=", 5) == 0) {
FLAGS_db = argv[i] + 5; FLAGS_db = argv[i] + 5;
} else { } else {
std::fprintf(stderr, "Invalid flag '%s'\n", argv[i]); fprintf(stderr, "Invalid flag '%s'\n", argv[i]);
std::exit(1); exit(1);
} }
} }
// Choose a location for the test database if none given with --db=<path> // Choose a location for the test database if none given with --db=<path>
if (FLAGS_db == nullptr) { if (FLAGS_db == NULL) {
leveldb::Env::Default()->GetTestDirectory(&default_db_path); leveldb::Env::Default()->GetTestDirectory(&default_db_path);
default_db_path += "/dbbench"; default_db_path += "/dbbench";
FLAGS_db = default_db_path.c_str(); FLAGS_db = default_db_path.c_str();
} }
leveldb::Benchmark benchmark; leveldb::Benchmark benchmark;

View File

@ -2,11 +2,9 @@
// Use of this source code is governed by a BSD-style license that can be // Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file. See the AUTHORS file for names of contributors. // found in the LICENSE file. See the AUTHORS file for names of contributors.
#include <stdio.h>
#include <stdlib.h>
#include <kcpolydb.h> #include <kcpolydb.h>
#include <cstdio>
#include <cstdlib>
#include "util/histogram.h" #include "util/histogram.h"
#include "util/random.h" #include "util/random.h"
#include "util/testutil.h" #include "util/testutil.h"
@ -36,7 +34,8 @@ static const char* FLAGS_benchmarks =
"fillrand100K," "fillrand100K,"
"fillseq100K," "fillseq100K,"
"readseq100K," "readseq100K,"
"readrand100K,"; "readrand100K,"
;
// Number of key/values to place in database // Number of key/values to place in database
static int FLAGS_num = 1000000; static int FLAGS_num = 1000000;
@ -70,12 +69,14 @@ static bool FLAGS_use_existing_db = false;
static bool FLAGS_compression = true; static bool FLAGS_compression = true;
// Use the db with the following name. // Use the db with the following name.
static const char* FLAGS_db = nullptr; static const char* FLAGS_db = NULL;
inline static void DBSynchronize(kyotocabinet::TreeDB* db_) { inline
static void DBSynchronize(kyotocabinet::TreeDB* db_)
{
// Synchronize will flush writes to disk // Synchronize will flush writes to disk
if (!db_->synchronize()) { if (!db_->synchronize()) {
std::fprintf(stderr, "synchronize error: %s\n", db_->error().name()); fprintf(stderr, "synchronize error: %s\n", db_->error().name());
} }
} }
@ -120,7 +121,7 @@ static Slice TrimSpace(Slice s) {
start++; start++;
} }
int limit = s.size(); int limit = s.size();
while (limit > start && isspace(s[limit - 1])) { while (limit > start && isspace(s[limit-1])) {
limit--; limit--;
} }
return Slice(s.data() + start, limit - start); return Slice(s.data() + start, limit - start);
@ -145,60 +146,55 @@ class Benchmark {
// State kept for progress messages // State kept for progress messages
int done_; int done_;
int next_report_; // When to report next int next_report_; // When to report next
void PrintHeader() { void PrintHeader() {
const int kKeySize = 16; const int kKeySize = 16;
PrintEnvironment(); PrintEnvironment();
std::fprintf(stdout, "Keys: %d bytes each\n", kKeySize); fprintf(stdout, "Keys: %d bytes each\n", kKeySize);
std::fprintf( fprintf(stdout, "Values: %d bytes each (%d bytes after compression)\n",
stdout, "Values: %d bytes each (%d bytes after compression)\n", FLAGS_value_size,
FLAGS_value_size, static_cast<int>(FLAGS_value_size * FLAGS_compression_ratio + 0.5));
static_cast<int>(FLAGS_value_size * FLAGS_compression_ratio + 0.5)); fprintf(stdout, "Entries: %d\n", num_);
std::fprintf(stdout, "Entries: %d\n", num_); fprintf(stdout, "RawSize: %.1f MB (estimated)\n",
std::fprintf(stdout, "RawSize: %.1f MB (estimated)\n", ((static_cast<int64_t>(kKeySize + FLAGS_value_size) * num_)
((static_cast<int64_t>(kKeySize + FLAGS_value_size) * num_) / / 1048576.0));
1048576.0)); fprintf(stdout, "FileSize: %.1f MB (estimated)\n",
std::fprintf( (((kKeySize + FLAGS_value_size * FLAGS_compression_ratio) * num_)
stdout, "FileSize: %.1f MB (estimated)\n", / 1048576.0));
(((kKeySize + FLAGS_value_size * FLAGS_compression_ratio) * num_) /
1048576.0));
PrintWarnings(); PrintWarnings();
std::fprintf(stdout, "------------------------------------------------\n"); fprintf(stdout, "------------------------------------------------\n");
} }
void PrintWarnings() { void PrintWarnings() {
#if defined(__GNUC__) && !defined(__OPTIMIZE__) #if defined(__GNUC__) && !defined(__OPTIMIZE__)
std::fprintf( fprintf(stdout,
stdout, "WARNING: Optimization is disabled: benchmarks unnecessarily slow\n"
"WARNING: Optimization is disabled: benchmarks unnecessarily slow\n"); );
#endif #endif
#ifndef NDEBUG #ifndef NDEBUG
std::fprintf( fprintf(stdout,
stdout, "WARNING: Assertions are enabled; benchmarks unnecessarily slow\n");
"WARNING: Assertions are enabled; benchmarks unnecessarily slow\n");
#endif #endif
} }
void PrintEnvironment() { void PrintEnvironment() {
std::fprintf( fprintf(stderr, "Kyoto Cabinet: version %s, lib ver %d, lib rev %d\n",
stderr, "Kyoto Cabinet: version %s, lib ver %d, lib rev %d\n", kyotocabinet::VERSION, kyotocabinet::LIBVER, kyotocabinet::LIBREV);
kyotocabinet::VERSION, kyotocabinet::LIBVER, kyotocabinet::LIBREV);
#if defined(__linux) #if defined(__linux)
time_t now = time(nullptr); time_t now = time(NULL);
std::fprintf(stderr, "Date: %s", fprintf(stderr, "Date: %s", ctime(&now)); // ctime() adds newline
ctime(&now)); // ctime() adds newline
FILE* cpuinfo = std::fopen("/proc/cpuinfo", "r"); FILE* cpuinfo = fopen("/proc/cpuinfo", "r");
if (cpuinfo != nullptr) { if (cpuinfo != NULL) {
char line[1000]; char line[1000];
int num_cpus = 0; int num_cpus = 0;
std::string cpu_type; std::string cpu_type;
std::string cache_size; std::string cache_size;
while (fgets(line, sizeof(line), cpuinfo) != nullptr) { while (fgets(line, sizeof(line), cpuinfo) != NULL) {
const char* sep = strchr(line, ':'); const char* sep = strchr(line, ':');
if (sep == nullptr) { if (sep == NULL) {
continue; continue;
} }
Slice key = TrimSpace(Slice(line, sep - 1 - line)); Slice key = TrimSpace(Slice(line, sep - 1 - line));
@ -210,10 +206,9 @@ class Benchmark {
cache_size = val.ToString(); cache_size = val.ToString();
} }
} }
std::fclose(cpuinfo); fclose(cpuinfo);
std::fprintf(stderr, "CPU: %d * %s\n", num_cpus, fprintf(stderr, "CPU: %d * %s\n", num_cpus, cpu_type.c_str());
cpu_type.c_str()); fprintf(stderr, "CPUCache: %s\n", cache_size.c_str());
std::fprintf(stderr, "CPUCache: %s\n", cache_size.c_str());
} }
#endif #endif
} }
@ -234,30 +229,23 @@ class Benchmark {
double micros = (now - last_op_finish_) * 1e6; double micros = (now - last_op_finish_) * 1e6;
hist_.Add(micros); hist_.Add(micros);
if (micros > 20000) { if (micros > 20000) {
std::fprintf(stderr, "long op: %.1f micros%30s\r", micros, ""); fprintf(stderr, "long op: %.1f micros%30s\r", micros, "");
std::fflush(stderr); fflush(stderr);
} }
last_op_finish_ = now; last_op_finish_ = now;
} }
done_++; done_++;
if (done_ >= next_report_) { if (done_ >= next_report_) {
if (next_report_ < 1000) if (next_report_ < 1000) next_report_ += 100;
next_report_ += 100; else if (next_report_ < 5000) next_report_ += 500;
else if (next_report_ < 5000) else if (next_report_ < 10000) next_report_ += 1000;
next_report_ += 500; else if (next_report_ < 50000) next_report_ += 5000;
else if (next_report_ < 10000) else if (next_report_ < 100000) next_report_ += 10000;
next_report_ += 1000; else if (next_report_ < 500000) next_report_ += 50000;
else if (next_report_ < 50000) else next_report_ += 100000;
next_report_ += 5000; fprintf(stderr, "... finished %d ops%30s\r", done_, "");
else if (next_report_ < 100000) fflush(stderr);
next_report_ += 10000;
else if (next_report_ < 500000)
next_report_ += 50000;
else
next_report_ += 100000;
std::fprintf(stderr, "... finished %d ops%30s\r", done_, "");
std::fflush(stderr);
} }
} }
@ -270,35 +258,42 @@ class Benchmark {
if (bytes_ > 0) { if (bytes_ > 0) {
char rate[100]; char rate[100];
std::snprintf(rate, sizeof(rate), "%6.1f MB/s", snprintf(rate, sizeof(rate), "%6.1f MB/s",
(bytes_ / 1048576.0) / (finish - start_)); (bytes_ / 1048576.0) / (finish - start_));
if (!message_.empty()) { if (!message_.empty()) {
message_ = std::string(rate) + " " + message_; message_ = std::string(rate) + " " + message_;
} else { } else {
message_ = rate; message_ = rate;
} }
} }
std::fprintf(stdout, "%-12s : %11.3f micros/op;%s%s\n", fprintf(stdout, "%-12s : %11.3f micros/op;%s%s\n",
name.ToString().c_str(), (finish - start_) * 1e6 / done_, name.ToString().c_str(),
(message_.empty() ? "" : " "), message_.c_str()); (finish - start_) * 1e6 / done_,
(message_.empty() ? "" : " "),
message_.c_str());
if (FLAGS_histogram) { if (FLAGS_histogram) {
std::fprintf(stdout, "Microseconds per op:\n%s\n", fprintf(stdout, "Microseconds per op:\n%s\n", hist_.ToString().c_str());
hist_.ToString().c_str());
} }
std::fflush(stdout); fflush(stdout);
} }
public: public:
enum Order { SEQUENTIAL, RANDOM }; enum Order {
enum DBState { FRESH, EXISTING }; SEQUENTIAL,
RANDOM
};
enum DBState {
FRESH,
EXISTING
};
Benchmark() Benchmark()
: db_(nullptr), : db_(NULL),
num_(FLAGS_num), num_(FLAGS_num),
reads_(FLAGS_reads < 0 ? FLAGS_num : FLAGS_reads), reads_(FLAGS_reads < 0 ? FLAGS_num : FLAGS_reads),
bytes_(0), bytes_(0),
rand_(301) { rand_(301) {
std::vector<std::string> files; std::vector<std::string> files;
std::string test_dir; std::string test_dir;
Env::Default()->GetTestDirectory(&test_dir); Env::Default()->GetTestDirectory(&test_dir);
@ -309,7 +304,7 @@ class Benchmark {
std::string file_name(test_dir); std::string file_name(test_dir);
file_name += "/"; file_name += "/";
file_name += files[i]; file_name += files[i];
Env::Default()->RemoveFile(file_name.c_str()); Env::Default()->DeleteFile(file_name.c_str());
} }
} }
} }
@ -317,7 +312,7 @@ class Benchmark {
~Benchmark() { ~Benchmark() {
if (!db_->close()) { if (!db_->close()) {
std::fprintf(stderr, "close error: %s\n", db_->error().name()); fprintf(stderr, "close error: %s\n", db_->error().name());
} }
} }
@ -326,12 +321,12 @@ class Benchmark {
Open(false); Open(false);
const char* benchmarks = FLAGS_benchmarks; const char* benchmarks = FLAGS_benchmarks;
while (benchmarks != nullptr) { while (benchmarks != NULL) {
const char* sep = strchr(benchmarks, ','); const char* sep = strchr(benchmarks, ',');
Slice name; Slice name;
if (sep == nullptr) { if (sep == NULL) {
name = benchmarks; name = benchmarks;
benchmarks = nullptr; benchmarks = NULL;
} else { } else {
name = Slice(benchmarks, sep - benchmarks); name = Slice(benchmarks, sep - benchmarks);
benchmarks = sep + 1; benchmarks = sep + 1;
@ -381,8 +376,7 @@ class Benchmark {
} else { } else {
known = false; known = false;
if (name != Slice()) { // No error message for empty name if (name != Slice()) { // No error message for empty name
std::fprintf(stderr, "unknown benchmark '%s'\n", fprintf(stderr, "unknown benchmark '%s'\n", name.ToString().c_str());
name.ToString().c_str());
} }
} }
if (known) { if (known) {
@ -392,8 +386,8 @@ class Benchmark {
} }
private: private:
void Open(bool sync) { void Open(bool sync) {
assert(db_ == nullptr); assert(db_ == NULL);
// Initialize db_ // Initialize db_
db_ = new kyotocabinet::TreeDB(); db_ = new kyotocabinet::TreeDB();
@ -401,14 +395,16 @@ class Benchmark {
db_num_++; db_num_++;
std::string test_dir; std::string test_dir;
Env::Default()->GetTestDirectory(&test_dir); Env::Default()->GetTestDirectory(&test_dir);
std::snprintf(file_name, sizeof(file_name), "%s/dbbench_polyDB-%d.kct", snprintf(file_name, sizeof(file_name),
test_dir.c_str(), db_num_); "%s/dbbench_polyDB-%d.kct",
test_dir.c_str(),
db_num_);
// Create tuning options and open the database // Create tuning options and open the database
int open_options = int open_options = kyotocabinet::PolyDB::OWRITER |
kyotocabinet::PolyDB::OWRITER | kyotocabinet::PolyDB::OCREATE; kyotocabinet::PolyDB::OCREATE;
int tune_options = int tune_options = kyotocabinet::TreeDB::TSMALL |
kyotocabinet::TreeDB::TSMALL | kyotocabinet::TreeDB::TLINEAR; kyotocabinet::TreeDB::TLINEAR;
if (FLAGS_compression) { if (FLAGS_compression) {
tune_options |= kyotocabinet::TreeDB::TCOMPRESS; tune_options |= kyotocabinet::TreeDB::TCOMPRESS;
db_->tune_compressor(&comp_); db_->tune_compressor(&comp_);
@ -416,17 +412,17 @@ class Benchmark {
db_->tune_options(tune_options); db_->tune_options(tune_options);
db_->tune_page_cache(FLAGS_cache_size); db_->tune_page_cache(FLAGS_cache_size);
db_->tune_page(FLAGS_page_size); db_->tune_page(FLAGS_page_size);
db_->tune_map(256LL << 20); db_->tune_map(256LL<<20);
if (sync) { if (sync) {
open_options |= kyotocabinet::PolyDB::OAUTOSYNC; open_options |= kyotocabinet::PolyDB::OAUTOSYNC;
} }
if (!db_->open(file_name, open_options)) { if (!db_->open(file_name, open_options)) {
std::fprintf(stderr, "open error: %s\n", db_->error().name()); fprintf(stderr, "open error: %s\n", db_->error().name());
} }
} }
void Write(bool sync, Order order, DBState state, int num_entries, void Write(bool sync, Order order, DBState state,
int value_size, int entries_per_batch) { int num_entries, int value_size, int entries_per_batch) {
// Create new database if state == FRESH // Create new database if state == FRESH
if (state == FRESH) { if (state == FRESH) {
if (FLAGS_use_existing_db) { if (FLAGS_use_existing_db) {
@ -434,26 +430,27 @@ class Benchmark {
return; return;
} }
delete db_; delete db_;
db_ = nullptr; db_ = NULL;
Open(sync); Open(sync);
Start(); // Do not count time taken to destroy/open Start(); // Do not count time taken to destroy/open
} }
if (num_entries != num_) { if (num_entries != num_) {
char msg[100]; char msg[100];
std::snprintf(msg, sizeof(msg), "(%d ops)", num_entries); snprintf(msg, sizeof(msg), "(%d ops)", num_entries);
message_ = msg; message_ = msg;
} }
// Write to database // Write to database
for (int i = 0; i < num_entries; i++) { for (int i = 0; i < num_entries; i++)
{
const int k = (order == SEQUENTIAL) ? i : (rand_.Next() % num_entries); const int k = (order == SEQUENTIAL) ? i : (rand_.Next() % num_entries);
char key[100]; char key[100];
std::snprintf(key, sizeof(key), "%016d", k); snprintf(key, sizeof(key), "%016d", k);
bytes_ += value_size + strlen(key); bytes_ += value_size + strlen(key);
std::string cpp_key = key; std::string cpp_key = key;
if (!db_->set(cpp_key, gen_.Generate(value_size).ToString())) { if (!db_->set(cpp_key, gen_.Generate(value_size).ToString())) {
std::fprintf(stderr, "set error: %s\n", db_->error().name()); fprintf(stderr, "set error: %s\n", db_->error().name());
} }
FinishedSingleOp(); FinishedSingleOp();
} }
@ -475,7 +472,7 @@ class Benchmark {
for (int i = 0; i < reads_; i++) { for (int i = 0; i < reads_; i++) {
char key[100]; char key[100];
const int k = rand_.Next() % reads_; const int k = rand_.Next() % reads_;
std::snprintf(key, sizeof(key), "%016d", k); snprintf(key, sizeof(key), "%016d", k);
db_->get(key, &value); db_->get(key, &value);
FinishedSingleOp(); FinishedSingleOp();
} }
@ -513,16 +510,16 @@ int main(int argc, char** argv) {
} else if (strncmp(argv[i], "--db=", 5) == 0) { } else if (strncmp(argv[i], "--db=", 5) == 0) {
FLAGS_db = argv[i] + 5; FLAGS_db = argv[i] + 5;
} else { } else {
std::fprintf(stderr, "Invalid flag '%s'\n", argv[i]); fprintf(stderr, "Invalid flag '%s'\n", argv[i]);
std::exit(1); exit(1);
} }
} }
// Choose a location for the test database if none given with --db=<path> // Choose a location for the test database if none given with --db=<path>
if (FLAGS_db == nullptr) { if (FLAGS_db == NULL) {
leveldb::Env::Default()->GetTestDirectory(&default_db_path); leveldb::Env::Default()->GetTestDirectory(&default_db_path);
default_db_path += "/dbbench"; default_db_path += "/dbbench";
FLAGS_db = default_db_path.c_str(); FLAGS_db = default_db_path.c_str();
} }
leveldb::Benchmark benchmark; leveldb::Benchmark benchmark;

View File

@ -83,23 +83,23 @@ div.bsql {
<p>Google, July 2011</p> <p>Google, July 2011</p>
<hr> <hr>
<p>In order to test LevelDB's performance, we benchmark it against other well-established database implementations. We compare LevelDB (revision 39) against <a href="https://www.sqlite.org/">SQLite3</a> (version 3.7.6.3) and <a href="https://dbmx.net/kyotocabinet/spex.html">Kyoto Cabinet's</a> (version 1.2.67) TreeDB (a B+Tree based key-value store). We would like to acknowledge Scott Hess and Mikio Hirabayashi for their suggestions and contributions to the SQLite3 and Kyoto Cabinet benchmarks, respectively.</p> <p>In order to test LevelDB's performance, we benchmark it against other well-established database implementations. We compare LevelDB (revision 39) against <a href="http://www.sqlite.org/">SQLite3</a> (version 3.7.6.3) and <a href="http://fallabs.com/kyotocabinet/spex.html">Kyoto Cabinet's</a> (version 1.2.67) TreeDB (a B+Tree based key-value store). We would like to acknowledge Scott Hess and Mikio Hirabayashi for their suggestions and contributions to the SQLite3 and Kyoto Cabinet benchmarks, respectively.</p>
<p>Benchmarks were all performed on a six-core Intel(R) Xeon(R) CPU X5650 @ 2.67GHz, with 12288 KB of total L3 cache and 12 GB of DDR3 RAM at 1333 MHz. (Note that LevelDB uses at most two CPUs since the benchmarks are single threaded: one to run the benchmark, and one for background compactions.) We ran the benchmarks on two machines (with identical processors), one with an Ext3 file system and one with an Ext4 file system. The machine with the Ext3 file system has a SATA Hitachi HDS721050CLA362 hard drive. The machine with the Ext4 file system has a SATA Samsung HD502HJ hard drive. Both hard drives spin at 7200 RPM and have hard drive write-caching enabled (using `hdparm -W 1 [device]`). The numbers reported below are the median of three measurements.</p> <p>Benchmarks were all performed on a six-core Intel(R) Xeon(R) CPU X5650 @ 2.67GHz, with 12288 KB of total L3 cache and 12 GB of DDR3 RAM at 1333 MHz. (Note that LevelDB uses at most two CPUs since the benchmarks are single threaded: one to run the benchmark, and one for background compactions.) We ran the benchmarks on two machines (with identical processors), one with an Ext3 file system and one with an Ext4 file system. The machine with the Ext3 file system has a SATA Hitachi HDS721050CLA362 hard drive. The machine with the Ext4 file system has a SATA Samsung HD502HJ hard drive. Both hard drives spin at 7200 RPM and have hard drive write-caching enabled (using `hdparm -W 1 [device]`). The numbers reported below are the median of three measurements.</p>
<h4>Benchmark Source Code</h4> <h4>Benchmark Source Code</h4>
<p>We wrote benchmark tools for SQLite and Kyoto TreeDB based on LevelDB's <span class="code">db_bench</span>. The code for each of the benchmarks resides here:</p> <p>We wrote benchmark tools for SQLite and Kyoto TreeDB based on LevelDB's <span class="code">db_bench</span>. The code for each of the benchmarks resides here:</p>
<ul> <ul>
<li> <b>LevelDB:</b> <a href="https://github.com/google/leveldb/blob/main/benchmarks/db_bench.cc">benchmarks/db_bench.cc</a>.</li> <li> <b>LevelDB:</b> <a href="http://code.google.com/p/leveldb/source/browse/trunk/db/db_bench.cc">db/db_bench.cc</a>.</li>
<li> <b>SQLite:</b> <a href="https://github.com/google/leveldb/blob/main/benchmarks/db_bench_sqlite3.cc">benchmarks/db_bench_sqlite3.cc</a>.</li> <li> <b>SQLite:</b> <a href="http://code.google.com/p/leveldb/source/browse/#svn%2Ftrunk%2Fdoc%2Fbench%2Fdb_bench_sqlite3.cc">doc/bench/db_bench_sqlite3.cc</a>.</li>
<li> <b>Kyoto TreeDB:</b> <a href="https://github.com/google/leveldb/blob/main/benchmarks/db_bench_tree_db.cc">benchmarks/db_bench_tree_db.cc</a>.</li> <li> <b>Kyoto TreeDB:</b> <a href="http://code.google.com/p/leveldb/source/browse/#svn%2Ftrunk%2Fdoc%2Fbench%2Fdb_bench_tree_db.cc">doc/bench/db_bench_tree_db.cc</a>.</li>
</ul> </ul>
<h4>Custom Build Specifications</h4> <h4>Custom Build Specifications</h4>
<ul> <ul>
<li>LevelDB: LevelDB was compiled with the <a href="https://github.com/gperftools/gperftools">tcmalloc</a> library and the <a href="https://github.com/google/snappy">Snappy</a> compression library (revision 33). Assertions were disabled.</li> <li>LevelDB: LevelDB was compiled with the <a href="http://code.google.com/p/google-perftools">tcmalloc</a> library and the <a href="http://code.google.com/p/snappy/">Snappy</a> compression library (revision 33). Assertions were disabled.</li>
<li>TreeDB: TreeDB was compiled using the <a href="https://www.oberhumer.com/opensource/lzo/">LZO</a> compression library (version 2.03). Furthermore, we enabled the TSMALL and TLINEAR options when opening the database in order to reduce the footprint of each record.</li> <li>TreeDB: TreeDB was compiled using the <a href="http://www.oberhumer.com/opensource/lzo/">LZO</a> compression library (version 2.03). Furthermore, we enabled the TSMALL and TLINEAR options when opening the database in order to reduce the footprint of each record.</li>
<li>SQLite: We tuned SQLite's performance, by setting its locking mode to exclusive. We also enabled SQLite's <a href="https://www.sqlite.org/draft/wal.html">write-ahead logging</a>.</li> <li>SQLite: We tuned SQLite's performance, by setting its locking mode to exclusive. We also enabled SQLite's <a href="http://www.sqlite.org/draft/wal.html">write-ahead logging</a>.</li>
</ul> </ul>
<h2>1. Baseline Performance</h2> <h2>1. Baseline Performance</h2>
@ -451,7 +451,7 @@ performance may very well be better with compression if it allows more
of the working set to fit in memory.</p> of the working set to fit in memory.</p>
<h2>Note about Ext4 Filesystems</h2> <h2>Note about Ext4 Filesystems</h2>
<p>The preceding numbers are for an ext3 file system. Synchronous writes are much slower under <a href="https://en.wikipedia.org/wiki/Ext4">ext4</a> (LevelDB drops to ~31 writes / second and TreeDB drops to ~5 writes / second; SQLite3's synchronous writes do not noticeably drop) due to ext4's different handling of <span class="code">fsync</span> / <span class="code">msync</span> calls. Even LevelDB's asynchronous write performance drops somewhat since it spreads its storage across multiple files and issues <span class="code">fsync</span> calls when switching to a new file.</p> <p>The preceding numbers are for an ext3 file system. Synchronous writes are much slower under <a href="http://en.wikipedia.org/wiki/Ext4">ext4</a> (LevelDB drops to ~31 writes / second and TreeDB drops to ~5 writes / second; SQLite3's synchronous writes do not noticeably drop) due to ext4's different handling of <span class="code">fsync</span> / <span class="code">msync</span> calls. Even LevelDB's asynchronous write performance drops somewhat since it spreads its storage across multiple files and issues <span class="code">fsync</span> calls when switching to a new file.</p>
<h2>Acknowledgements</h2> <h2>Acknowledgements</h2>
<p>Jeff Dean and Sanjay Ghemawat wrote LevelDB. Kevin Tseng wrote and compiled these benchmarks. Mikio Hirabayashi, Scott Hess, and Gabor Cselle provided help and advice.</p> <p>Jeff Dean and Sanjay Ghemawat wrote LevelDB. Kevin Tseng wrote and compiled these benchmarks. Mikio Hirabayashi, Scott Hess, and Gabor Cselle provided help and advice.</p>

89
doc/doc.css Normal file
View File

@ -0,0 +1,89 @@
body {
margin-left: 0.5in;
margin-right: 0.5in;
background: white;
color: black;
}
h1 {
margin-left: -0.2in;
font-size: 14pt;
}
h2 {
margin-left: -0in;
font-size: 12pt;
}
h3 {
margin-left: -0in;
}
h4 {
margin-left: -0in;
}
hr {
margin-left: -0in;
}
/* Definition lists: definition term bold */
dt {
font-weight: bold;
}
address {
text-align: center;
}
code,samp,var {
color: blue;
}
kbd {
color: #600000;
}
div.note p {
float: right;
width: 3in;
margin-right: 0%;
padding: 1px;
border: 2px solid #6060a0;
background-color: #fffff0;
}
ul {
margin-top: -0em;
margin-bottom: -0em;
}
ol {
margin-top: -0em;
margin-bottom: -0em;
}
UL.nobullets {
list-style-type: none;
list-style-image: none;
margin-left: -1em;
}
p {
margin: 1em 0 1em 0;
padding: 0 0 0 0;
}
pre {
line-height: 1.3em;
padding: 0.4em 0 0.8em 0;
margin: 0 0 0 0;
border: 0 0 0 0;
color: blue;
}
.datatable {
margin-left: auto;
margin-right: auto;
margin-top: 2em;
margin-bottom: 2em;
border: 1px solid;
}
.datatable td,th {
padding: 0 0.5em 0 0.5em;
text-align: right;
}

213
doc/impl.html Normal file
View File

@ -0,0 +1,213 @@
<!DOCTYPE html>
<html>
<head>
<link rel="stylesheet" type="text/css" href="doc.css" />
<title>Leveldb file layout and compactions</title>
</head>
<body>
<h1>Files</h1>
The implementation of leveldb is similar in spirit to the
representation of a single
<a href="http://research.google.com/archive/bigtable.html">
Bigtable tablet (section 5.3)</a>.
However the organization of the files that make up the representation
is somewhat different and is explained below.
<p>
Each database is represented by a set of files stored in a directory.
There are several different types of files as documented below:
<p>
<h2>Log files</h2>
<p>
A log file (*.log) stores a sequence of recent updates. Each update
is appended to the current log file. When the log file reaches a
pre-determined size (approximately 4MB by default), it is converted
to a sorted table (see below) and a new log file is created for future
updates.
<p>
A copy of the current log file is kept in an in-memory structure (the
<code>memtable</code>). This copy is consulted on every read so that read
operations reflect all logged updates.
<p>
<h2>Sorted tables</h2>
<p>
A sorted table (*.sst) stores a sequence of entries sorted by key.
Each entry is either a value for the key, or a deletion marker for the
key. (Deletion markers are kept around to hide obsolete values
present in older sorted tables).
<p>
The set of sorted tables are organized into a sequence of levels. The
sorted table generated from a log file is placed in a special <code>young</code>
level (also called level-0). When the number of young files exceeds a
certain threshold (currently four), all of the young files are merged
together with all of the overlapping level-1 files to produce a
sequence of new level-1 files (we create a new level-1 file for every
2MB of data.)
<p>
Files in the young level may contain overlapping keys. However files
in other levels have distinct non-overlapping key ranges. Consider
level number L where L >= 1. When the combined size of files in
level-L exceeds (10^L) MB (i.e., 10MB for level-1, 100MB for level-2,
...), one file in level-L, and all of the overlapping files in
level-(L+1) are merged to form a set of new files for level-(L+1).
These merges have the effect of gradually migrating new updates from
the young level to the largest level using only bulk reads and writes
(i.e., minimizing expensive seeks).
<h2>Manifest</h2>
<p>
A MANIFEST file lists the set of sorted tables that make up each
level, the corresponding key ranges, and other important metadata.
A new MANIFEST file (with a new number embedded in the file name)
is created whenever the database is reopened. The MANIFEST file is
formatted as a log, and changes made to the serving state (as files
are added or removed) are appended to this log.
<p>
<h2>Current</h2>
<p>
CURRENT is a simple text file that contains the name of the latest
MANIFEST file.
<p>
<h2>Info logs</h2>
<p>
Informational messages are printed to files named LOG and LOG.old.
<p>
<h2>Others</h2>
<p>
Other files used for miscellaneous purposes may also be present
(LOCK, *.dbtmp).
<h1>Level 0</h1>
When the log file grows above a certain size (1MB by default):
<ul>
<li>Create a brand new memtable and log file and direct future updates here
<li>In the background:
<ul>
<li>Write the contents of the previous memtable to an sstable
<li>Discard the memtable
<li>Delete the old log file and the old memtable
<li>Add the new sstable to the young (level-0) level.
</ul>
</ul>
<h1>Compactions</h1>
<p>
When the size of level L exceeds its limit, we compact it in a
background thread. The compaction picks a file from level L and all
overlapping files from the next level L+1. Note that if a level-L
file overlaps only part of a level-(L+1) file, the entire file at
level-(L+1) is used as an input to the compaction and will be
discarded after the compaction. Aside: because level-0 is special
(files in it may overlap each other), we treat compactions from
level-0 to level-1 specially: a level-0 compaction may pick more than
one level-0 file in case some of these files overlap each other.
<p>
A compaction merges the contents of the picked files to produce a
sequence of level-(L+1) files. We switch to producing a new
level-(L+1) file after the current output file has reached the target
file size (2MB). We also switch to a new output file when the key
range of the current output file has grown enough to overlap more than
ten level-(L+2) files. This last rule ensures that a later compaction
of a level-(L+1) file will not pick up too much data from level-(L+2).
<p>
The old files are discarded and the new files are added to the serving
state.
<p>
Compactions for a particular level rotate through the key space. In
more detail, for each level L, we remember the ending key of the last
compaction at level L. The next compaction for level L will pick the
first file that starts after this key (wrapping around to the
beginning of the key space if there is no such file).
<p>
Compactions drop overwritten values. They also drop deletion markers
if there are no higher numbered levels that contain a file whose range
overlaps the current key.
<h2>Timing</h2>
Level-0 compactions will read up to four 1MB files from level-0, and
at worst all the level-1 files (10MB). I.e., we will read 14MB and
write 14MB.
<p>
Other than the special level-0 compactions, we will pick one 2MB file
from level L. In the worst case, this will overlap ~ 12 files from
level L+1 (10 because level-(L+1) is ten times the size of level-L,
and another two at the boundaries since the file ranges at level-L
will usually not be aligned with the file ranges at level-L+1). The
compaction will therefore read 26MB and write 26MB. Assuming a disk
IO rate of 100MB/s (ballpark range for modern drives), the worst
compaction cost will be approximately 0.5 second.
<p>
If we throttle the background writing to something small, say 10% of
the full 100MB/s speed, a compaction may take up to 5 seconds. If the
user is writing at 10MB/s, we might build up lots of level-0 files
(~50 to hold the 5*10MB). This may significantly increase the cost of
reads due to the overhead of merging more files together on every
read.
<p>
Solution 1: To reduce this problem, we might want to increase the log
switching threshold when the number of level-0 files is large. Though
the downside is that the larger this threshold, the more memory we will
need to hold the corresponding memtable.
<p>
Solution 2: We might want to decrease write rate artificially when the
number of level-0 files goes up.
<p>
Solution 3: We work on reducing the cost of very wide merges.
Perhaps most of the level-0 files will have their blocks sitting
uncompressed in the cache and we will only need to worry about the
O(N) complexity in the merging iterator.
<h2>Number of files</h2>
Instead of always making 2MB files, we could make larger files for
larger levels to reduce the total file count, though at the expense of
more bursty compactions. Alternatively, we could shard the set of
files into multiple directories.
<p>
An experiment on an <code>ext3</code> filesystem on Feb 04, 2011 shows
the following timings to do 100K file opens in directories with
varying number of files:
<table class="datatable">
<tr><th>Files in directory</th><th>Microseconds to open a file</th></tr>
<tr><td>1000</td><td>9</td>
<tr><td>10000</td><td>10</td>
<tr><td>100000</td><td>16</td>
</table>
So maybe even the sharding is not necessary on modern filesystems?
<h1>Recovery</h1>
<ul>
<li> Read CURRENT to find name of the latest committed MANIFEST
<li> Read the named MANIFEST file
<li> Clean up stale files
<li> We could open all sstables here, but it is probably better to be lazy...
<li> Convert log chunk to a new level-0 sstable
<li> Start directing new writes to a new log file with recovered sequence#
</ul>
<h1>Garbage collection of files</h1>
<code>DeleteObsoleteFiles()</code> is called at the end of every
compaction and at the end of recovery. It finds the names of all
files in the database. It deletes all log files that are not the
current log file. It deletes all table files that are not referenced
from some level and are not the output of an active compaction.
</body>
</html>

View File

@ -1,172 +0,0 @@
## Files
The implementation of leveldb is similar in spirit to the representation of a
single [Bigtable tablet (section 5.3)](https://research.google/pubs/pub27898/).
However the organization of the files that make up the representation is
somewhat different and is explained below.
Each database is represented by a set of files stored in a directory. There are
several different types of files as documented below:
### Log files
A log file (*.log) stores a sequence of recent updates. Each update is appended
to the current log file. When the log file reaches a pre-determined size
(approximately 4MB by default), it is converted to a sorted table (see below)
and a new log file is created for future updates.
A copy of the current log file is kept in an in-memory structure (the
`memtable`). This copy is consulted on every read so that read operations
reflect all logged updates.
## Sorted tables
A sorted table (*.ldb) stores a sequence of entries sorted by key. Each entry is
either a value for the key, or a deletion marker for the key. (Deletion markers
are kept around to hide obsolete values present in older sorted tables).
The set of sorted tables are organized into a sequence of levels. The sorted
table generated from a log file is placed in a special **young** level (also
called level-0). When the number of young files exceeds a certain threshold
(currently four), all of the young files are merged together with all of the
overlapping level-1 files to produce a sequence of new level-1 files (we create
a new level-1 file for every 2MB of data.)
Files in the young level may contain overlapping keys. However files in other
levels have distinct non-overlapping key ranges. Consider level number L where
L >= 1. When the combined size of files in level-L exceeds (10^L) MB (i.e., 10MB
for level-1, 100MB for level-2, ...), one file in level-L, and all of the
overlapping files in level-(L+1) are merged to form a set of new files for
level-(L+1). These merges have the effect of gradually migrating new updates
from the young level to the largest level using only bulk reads and writes
(i.e., minimizing expensive seeks).
### Manifest
A MANIFEST file lists the set of sorted tables that make up each level, the
corresponding key ranges, and other important metadata. A new MANIFEST file
(with a new number embedded in the file name) is created whenever the database
is reopened. The MANIFEST file is formatted as a log, and changes made to the
serving state (as files are added or removed) are appended to this log.
### Current
CURRENT is a simple text file that contains the name of the latest MANIFEST
file.
### Info logs
Informational messages are printed to files named LOG and LOG.old.
### Others
Other files used for miscellaneous purposes may also be present (LOCK, *.dbtmp).
## Level 0
When the log file grows above a certain size (4MB by default):
Create a brand new memtable and log file and direct future updates here.
In the background:
1. Write the contents of the previous memtable to an sstable.
2. Discard the memtable.
3. Delete the old log file and the old memtable.
4. Add the new sstable to the young (level-0) level.
## Compactions
When the size of level L exceeds its limit, we compact it in a background
thread. The compaction picks a file from level L and all overlapping files from
the next level L+1. Note that if a level-L file overlaps only part of a
level-(L+1) file, the entire file at level-(L+1) is used as an input to the
compaction and will be discarded after the compaction. Aside: because level-0
is special (files in it may overlap each other), we treat compactions from
level-0 to level-1 specially: a level-0 compaction may pick more than one
level-0 file in case some of these files overlap each other.
A compaction merges the contents of the picked files to produce a sequence of
level-(L+1) files. We switch to producing a new level-(L+1) file after the
current output file has reached the target file size (2MB). We also switch to a
new output file when the key range of the current output file has grown enough
to overlap more than ten level-(L+2) files. This last rule ensures that a later
compaction of a level-(L+1) file will not pick up too much data from
level-(L+2).
The old files are discarded and the new files are added to the serving state.
Compactions for a particular level rotate through the key space. In more detail,
for each level L, we remember the ending key of the last compaction at level L.
The next compaction for level L will pick the first file that starts after this
key (wrapping around to the beginning of the key space if there is no such
file).
Compactions drop overwritten values. They also drop deletion markers if there
are no higher numbered levels that contain a file whose range overlaps the
current key.
### Timing
Level-0 compactions will read up to four 1MB files from level-0, and at worst
all the level-1 files (10MB). I.e., we will read 14MB and write 14MB.
Other than the special level-0 compactions, we will pick one 2MB file from level
L. In the worst case, this will overlap ~ 12 files from level L+1 (10 because
level-(L+1) is ten times the size of level-L, and another two at the boundaries
since the file ranges at level-L will usually not be aligned with the file
ranges at level-L+1). The compaction will therefore read 26MB and write 26MB.
Assuming a disk IO rate of 100MB/s (ballpark range for modern drives), the worst
compaction cost will be approximately 0.5 second.
If we throttle the background writing to something small, say 10% of the full
100MB/s speed, a compaction may take up to 5 seconds. If the user is writing at
10MB/s, we might build up lots of level-0 files (~50 to hold the 5*10MB). This
may significantly increase the cost of reads due to the overhead of merging more
files together on every read.
Solution 1: To reduce this problem, we might want to increase the log switching
threshold when the number of level-0 files is large. Though the downside is that
the larger this threshold, the more memory we will need to hold the
corresponding memtable.
Solution 2: We might want to decrease write rate artificially when the number of
level-0 files goes up.
Solution 3: We work on reducing the cost of very wide merges. Perhaps most of
the level-0 files will have their blocks sitting uncompressed in the cache and
we will only need to worry about the O(N) complexity in the merging iterator.
### Number of files
Instead of always making 2MB files, we could make larger files for larger levels
to reduce the total file count, though at the expense of more bursty
compactions. Alternatively, we could shard the set of files into multiple
directories.
An experiment on an ext3 filesystem on Feb 04, 2011 shows the following timings
to do 100K file opens in directories with varying number of files:
| Files in directory | Microseconds to open a file |
|-------------------:|----------------------------:|
| 1000 | 9 |
| 10000 | 10 |
| 100000 | 16 |
So maybe even the sharding is not necessary on modern filesystems?
## Recovery
* Read CURRENT to find name of the latest committed MANIFEST
* Read the named MANIFEST file
* Clean up stale files
* We could open all sstables here, but it is probably better to be lazy...
* Convert log chunk to a new level-0 sstable
* Start directing new writes to a new log file with recovered sequence#
## Garbage collection of files
`RemoveObsoleteFiles()` is called at the end of every compaction and at the end
of recovery. It finds the names of all files in the database. It deletes all log
files that are not the current log file. It deletes all table files that are not
referenced from some level and are not the output of an active compaction.

549
doc/index.html Normal file
View File

@ -0,0 +1,549 @@
<!DOCTYPE html>
<html>
<head>
<link rel="stylesheet" type="text/css" href="doc.css" />
<title>Leveldb</title>
</head>
<body>
<h1>Leveldb</h1>
<address>Jeff Dean, Sanjay Ghemawat</address>
<p>
The <code>leveldb</code> library provides a persistent key value store. Keys and
values are arbitrary byte arrays. The keys are ordered within the key
value store according to a user-specified comparator function.
<p>
<h1>Opening A Database</h1>
<p>
A <code>leveldb</code> database has a name which corresponds to a file system
directory. All of the contents of database are stored in this
directory. The following example shows how to open a database,
creating it if necessary:
<p>
<pre>
#include &lt;assert&gt;
#include "leveldb/db.h"
leveldb::DB* db;
leveldb::Options options;
options.create_if_missing = true;
leveldb::Status status = leveldb::DB::Open(options, "/tmp/testdb", &amp;db);
assert(status.ok());
...
</pre>
If you want to raise an error if the database already exists, add
the following line before the <code>leveldb::DB::Open</code> call:
<pre>
options.error_if_exists = true;
</pre>
<h1>Status</h1>
<p>
You may have noticed the <code>leveldb::Status</code> type above. Values of this
type are returned by most functions in <code>leveldb</code> that may encounter an
error. You can check if such a result is ok, and also print an
associated error message:
<p>
<pre>
leveldb::Status s = ...;
if (!s.ok()) cerr &lt;&lt; s.ToString() &lt;&lt; endl;
</pre>
<h1>Closing A Database</h1>
<p>
When you are done with a database, just delete the database object.
Example:
<p>
<pre>
... open the db as described above ...
... do something with db ...
delete db;
</pre>
<h1>Reads And Writes</h1>
<p>
The database provides <code>Put</code>, <code>Delete</code>, and <code>Get</code> methods to
modify/query the database. For example, the following code
moves the value stored under key1 to key2.
<pre>
std::string value;
leveldb::Status s = db-&gt;Get(leveldb::ReadOptions(), key1, &amp;value);
if (s.ok()) s = db-&gt;Put(leveldb::WriteOptions(), key2, value);
if (s.ok()) s = db-&gt;Delete(leveldb::WriteOptions(), key1);
</pre>
<h1>Atomic Updates</h1>
<p>
Note that if the process dies after the Put of key2 but before the
delete of key1, the same value may be left stored under multiple keys.
Such problems can be avoided by using the <code>WriteBatch</code> class to
atomically apply a set of updates:
<p>
<pre>
#include "leveldb/write_batch.h"
...
std::string value;
leveldb::Status s = db-&gt;Get(leveldb::ReadOptions(), key1, &amp;value);
if (s.ok()) {
leveldb::WriteBatch batch;
batch.Delete(key1);
batch.Put(key2, value);
s = db-&gt;Write(leveldb::WriteOptions(), &amp;batch);
}
</pre>
The <code>WriteBatch</code> holds a sequence of edits to be made to the database,
and these edits within the batch are applied in order. Note that we
called <code>Delete</code> before <code>Put</code> so that if <code>key1</code> is identical to <code>key2</code>,
we do not end up erroneously dropping the value entirely.
<p>
Apart from its atomicity benefits, <code>WriteBatch</code> may also be used to
speed up bulk updates by placing lots of individual mutations into the
same batch.
<h1>Synchronous Writes</h1>
By default, each write to <code>leveldb</code> is asynchronous: it
returns after pushing the write from the process into the operating
system. The transfer from operating system memory to the underlying
persistent storage happens asynchronously. The <code>sync</code> flag
can be turned on for a particular write to make the write operation
not return until the data being written has been pushed all the way to
persistent storage. (On Posix systems, this is implemented by calling
either <code>fsync(...)</code> or <code>fdatasync(...)</code> or
<code>msync(..., MS_SYNC)</code> before the write operation returns.)
<pre>
leveldb::WriteOptions write_options;
write_options.sync = true;
db-&gt;Put(write_options, ...);
</pre>
Asynchronous writes are often more than a thousand times as fast as
synchronous writes. The downside of asynchronous writes is that a
crash of the machine may cause the last few updates to be lost. Note
that a crash of just the writing process (i.e., not a reboot) will not
cause any loss since even when <code>sync</code> is false, an update
is pushed from the process memory into the operating system before it
is considered done.
<p>
Asynchronous writes can often be used safely. For example, when
loading a large amount of data into the database you can handle lost
updates by restarting the bulk load after a crash. A hybrid scheme is
also possible where every Nth write is synchronous, and in the event
of a crash, the bulk load is restarted just after the last synchronous
write finished by the previous run. (The synchronous write can update
a marker that describes where to restart on a crash.)
<p>
<code>WriteBatch</code> provides an alternative to asynchronous writes.
Multiple updates may be placed in the same <code>WriteBatch</code> and
applied together using a synchronous write (i.e.,
<code>write_options.sync</code> is set to true). The extra cost of
the synchronous write will be amortized across all of the writes in
the batch.
<p>
<h1>Concurrency</h1>
<p>
A database may only be opened by one process at a time.
The <code>leveldb</code> implementation acquires a lock from the
operating system to prevent misuse. Within a single process, the
same <code>leveldb::DB</code> object may be safely shared by multiple
concurrent threads. I.e., different threads may write into or fetch
iterators or call <code>Get</code> on the same database without any
external synchronization (the leveldb implementation will
automatically do the required synchronization). However other objects
(like Iterator and WriteBatch) may require external synchronization.
If two threads share such an object, they must protect access to it
using their own locking protocol. More details are available in
the public header files.
<p>
<h1>Iteration</h1>
<p>
The following example demonstrates how to print all key,value pairs
in a database.
<p>
<pre>
leveldb::Iterator* it = db-&gt;NewIterator(leveldb::ReadOptions());
for (it-&gt;SeekToFirst(); it-&gt;Valid(); it-&gt;Next()) {
cout &lt;&lt; it-&gt;key().ToString() &lt;&lt; ": " &lt;&lt; it-&gt;value().ToString() &lt;&lt; endl;
}
assert(it-&gt;status().ok()); // Check for any errors found during the scan
delete it;
</pre>
The following variation shows how to process just the keys in the
range <code>[start,limit)</code>:
<p>
<pre>
for (it-&gt;Seek(start);
it-&gt;Valid() &amp;&amp; it-&gt;key().ToString() &lt; limit;
it-&gt;Next()) {
...
}
</pre>
You can also process entries in reverse order. (Caveat: reverse
iteration may be somewhat slower than forward iteration.)
<p>
<pre>
for (it-&gt;SeekToLast(); it-&gt;Valid(); it-&gt;Prev()) {
...
}
</pre>
<h1>Snapshots</h1>
<p>
Snapshots provide consistent read-only views over the entire state of
the key-value store. <code>ReadOptions::snapshot</code> may be non-NULL to indicate
that a read should operate on a particular version of the DB state.
If <code>ReadOptions::snapshot</code> is NULL, the read will operate on an
implicit snapshot of the current state.
<p>
Snapshots are created by the DB::GetSnapshot() method:
<p>
<pre>
leveldb::ReadOptions options;
options.snapshot = db-&gt;GetSnapshot();
... apply some updates to db ...
leveldb::Iterator* iter = db-&gt;NewIterator(options);
... read using iter to view the state when the snapshot was created ...
delete iter;
db-&gt;ReleaseSnapshot(options.snapshot);
</pre>
Note that when a snapshot is no longer needed, it should be released
using the DB::ReleaseSnapshot interface. This allows the
implementation to get rid of state that was being maintained just to
support reading as of that snapshot.
<h1>Slice</h1>
<p>
The return value of the <code>it->key()</code> and <code>it->value()</code> calls above
are instances of the <code>leveldb::Slice</code> type. <code>Slice</code> is a simple
structure that contains a length and a pointer to an external byte
array. Returning a <code>Slice</code> is a cheaper alternative to returning a
<code>std::string</code> since we do not need to copy potentially large keys and
values. In addition, <code>leveldb</code> methods do not return null-terminated
C-style strings since <code>leveldb</code> keys and values are allowed to
contain '\0' bytes.
<p>
C++ strings and null-terminated C-style strings can be easily converted
to a Slice:
<p>
<pre>
leveldb::Slice s1 = "hello";
std::string str("world");
leveldb::Slice s2 = str;
</pre>
A Slice can be easily converted back to a C++ string:
<pre>
std::string str = s1.ToString();
assert(str == std::string("hello"));
</pre>
Be careful when using Slices since it is up to the caller to ensure that
the external byte array into which the Slice points remains live while
the Slice is in use. For example, the following is buggy:
<p>
<pre>
leveldb::Slice slice;
if (...) {
std::string str = ...;
slice = str;
}
Use(slice);
</pre>
When the <code>if</code> statement goes out of scope, <code>str</code> will be destroyed and the
backing storage for <code>slice</code> will disappear.
<p>
<h1>Comparators</h1>
<p>
The preceding examples used the default ordering function for key,
which orders bytes lexicographically. You can however supply a custom
comparator when opening a database. For example, suppose each
database key consists of two numbers and we should sort by the first
number, breaking ties by the second number. First, define a proper
subclass of <code>leveldb::Comparator</code> that expresses these rules:
<p>
<pre>
class TwoPartComparator : public leveldb::Comparator {
public:
// Three-way comparison function:
// if a &lt; b: negative result
// if a &gt; b: positive result
// else: zero result
int Compare(const leveldb::Slice&amp; a, const leveldb::Slice&amp; b) const {
int a1, a2, b1, b2;
ParseKey(a, &amp;a1, &amp;a2);
ParseKey(b, &amp;b1, &amp;b2);
if (a1 &lt; b1) return -1;
if (a1 &gt; b1) return +1;
if (a2 &lt; b2) return -1;
if (a2 &gt; b2) return +1;
return 0;
}
// Ignore the following methods for now:
const char* Name() const { return "TwoPartComparator"; }
void FindShortestSeparator(std::string*, const leveldb::Slice&amp;) const { }
void FindShortSuccessor(std::string*) const { }
};
</pre>
Now create a database using this custom comparator:
<p>
<pre>
TwoPartComparator cmp;
leveldb::DB* db;
leveldb::Options options;
options.create_if_missing = true;
options.comparator = &amp;cmp;
leveldb::Status status = leveldb::DB::Open(options, "/tmp/testdb", &amp;db);
...
</pre>
<h2>Backwards compatibility</h2>
<p>
The result of the comparator's <code>Name</code> method is attached to the
database when it is created, and is checked on every subsequent
database open. If the name changes, the <code>leveldb::DB::Open</code> call will
fail. Therefore, change the name if and only if the new key format
and comparison function are incompatible with existing databases, and
it is ok to discard the contents of all existing databases.
<p>
You can however still gradually evolve your key format over time with
a little bit of pre-planning. For example, you could store a version
number at the end of each key (one byte should suffice for most uses).
When you wish to switch to a new key format (e.g., adding an optional
third part to the keys processed by <code>TwoPartComparator</code>),
(a) keep the same comparator name (b) increment the version number
for new keys (c) change the comparator function so it uses the
version numbers found in the keys to decide how to interpret them.
<p>
<h1>Performance</h1>
<p>
Performance can be tuned by changing the default values of the
types defined in <code>include/leveldb/options.h</code>.
<p>
<h2>Block size</h2>
<p>
<code>leveldb</code> groups adjacent keys together into the same block and such a
block is the unit of transfer to and from persistent storage. The
default block size is approximately 4096 uncompressed bytes.
Applications that mostly do bulk scans over the contents of the
database may wish to increase this size. Applications that do a lot
of point reads of small values may wish to switch to a smaller block
size if performance measurements indicate an improvement. There isn't
much benefit in using blocks smaller than one kilobyte, or larger than
a few megabytes. Also note that compression will be more effective
with larger block sizes.
<p>
<h2>Compression</h2>
<p>
Each block is individually compressed before being written to
persistent storage. Compression is on by default since the default
compression method is very fast, and is automatically disabled for
uncompressible data. In rare cases, applications may want to disable
compression entirely, but should only do so if benchmarks show a
performance improvement:
<p>
<pre>
leveldb::Options options;
options.compression = leveldb::kNoCompression;
... leveldb::DB::Open(options, name, ...) ....
</pre>
<h2>Cache</h2>
<p>
The contents of the database are stored in a set of files in the
filesystem and each file stores a sequence of compressed blocks. If
<code>options.cache</code> is non-NULL, it is used to cache frequently used
uncompressed block contents.
<p>
<pre>
#include "leveldb/cache.h"
leveldb::Options options;
options.cache = leveldb::NewLRUCache(100 * 1048576); // 100MB cache
leveldb::DB* db;
leveldb::DB::Open(options, name, &db);
... use the db ...
delete db
delete options.cache;
</pre>
Note that the cache holds uncompressed data, and therefore it should
be sized according to application level data sizes, without any
reduction from compression. (Caching of compressed blocks is left to
the operating system buffer cache, or any custom <code>Env</code>
implementation provided by the client.)
<p>
When performing a bulk read, the application may wish to disable
caching so that the data processed by the bulk read does not end up
displacing most of the cached contents. A per-iterator option can be
used to achieve this:
<p>
<pre>
leveldb::ReadOptions options;
options.fill_cache = false;
leveldb::Iterator* it = db-&gt;NewIterator(options);
for (it-&gt;SeekToFirst(); it-&gt;Valid(); it-&gt;Next()) {
...
}
</pre>
<h2>Key Layout</h2>
<p>
Note that the unit of disk transfer and caching is a block. Adjacent
keys (according to the database sort order) will usually be placed in
the same block. Therefore the application can improve its performance
by placing keys that are accessed together near each other and placing
infrequently used keys in a separate region of the key space.
<p>
For example, suppose we are implementing a simple file system on top
of <code>leveldb</code>. The types of entries we might wish to store are:
<p>
<pre>
filename -&gt; permission-bits, length, list of file_block_ids
file_block_id -&gt; data
</pre>
We might want to prefix <code>filename</code> keys with one letter (say '/') and the
<code>file_block_id</code> keys with a different letter (say '0') so that scans
over just the metadata do not force us to fetch and cache bulky file
contents.
<p>
<h2>Filters</h2>
<p>
Because of the way <code>leveldb</code> data is organized on disk,
a single <code>Get()</code> call may involve multiple reads from disk.
The optional <code>FilterPolicy</code> mechanism can be used to reduce
the number of disk reads substantially.
<pre>
leveldb::Options options;
options.filter_policy = NewBloomFilterPolicy(10);
leveldb::DB* db;
leveldb::DB::Open(options, "/tmp/testdb", &amp;db);
... use the database ...
delete db;
delete options.filter_policy;
</pre>
The preceding code associates a
<a href="http://en.wikipedia.org/wiki/Bloom_filter">Bloom filter</a>
based filtering policy with the database. Bloom filter based
filtering relies on keeping some number of bits of data in memory per
key (in this case 10 bits per key since that is the argument we passed
to NewBloomFilterPolicy). This filter will reduce the number of unnecessary
disk reads needed for <code>Get()</code> calls by a factor of
approximately a 100. Increasing the bits per key will lead to a
larger reduction at the cost of more memory usage. We recommend that
applications whose working set does not fit in memory and that do a
lot of random reads set a filter policy.
<p>
If you are using a custom comparator, you should ensure that the filter
policy you are using is compatible with your comparator. For example,
consider a comparator that ignores trailing spaces when comparing keys.
<code>NewBloomFilterPolicy</code> must not be used with such a comparator.
Instead, the application should provide a custom filter policy that
also ignores trailing spaces. For example:
<pre>
class CustomFilterPolicy : public leveldb::FilterPolicy {
private:
FilterPolicy* builtin_policy_;
public:
CustomFilterPolicy() : builtin_policy_(NewBloomFilterPolicy(10)) { }
~CustomFilterPolicy() { delete builtin_policy_; }
const char* Name() const { return "IgnoreTrailingSpacesFilter"; }
void CreateFilter(const Slice* keys, int n, std::string* dst) const {
// Use builtin bloom filter code after removing trailing spaces
std::vector&lt;Slice&gt; trimmed(n);
for (int i = 0; i &lt; n; i++) {
trimmed[i] = RemoveTrailingSpaces(keys[i]);
}
return builtin_policy_-&gt;CreateFilter(&amp;trimmed[i], n, dst);
}
bool KeyMayMatch(const Slice& key, const Slice& filter) const {
// Use builtin bloom filter code after removing trailing spaces
return builtin_policy_-&gt;KeyMayMatch(RemoveTrailingSpaces(key), filter);
}
};
</pre>
<p>
Advanced applications may provide a filter policy that does not use
a bloom filter but uses some other mechanism for summarizing a set
of keys. See <code>leveldb/filter_policy.h</code> for detail.
<p>
<h1>Checksums</h1>
<p>
<code>leveldb</code> associates checksums with all data it stores in the file system.
There are two separate controls provided over how aggressively these
checksums are verified:
<p>
<ul>
<li> <code>ReadOptions::verify_checksums</code> may be set to true to force
checksum verification of all data that is read from the file system on
behalf of a particular read. By default, no such verification is
done.
<p>
<li> <code>Options::paranoid_checks</code> may be set to true before opening a
database to make the database implementation raise an error as soon as
it detects an internal corruption. Depending on which portion of the
database has been corrupted, the error may be raised when the database
is opened, or later by another database operation. By default,
paranoid checking is off so that the database can be used even if
parts of its persistent storage have been corrupted.
<p>
If a database is corrupted (perhaps it cannot be opened when
paranoid checking is turned on), the <code>leveldb::RepairDB</code> function
may be used to recover as much of the data as possible
<p>
</ul>
<h1>Approximate Sizes</h1>
<p>
The <code>GetApproximateSizes</code> method can used to get the approximate
number of bytes of file system space used by one or more key ranges.
<p>
<pre>
leveldb::Range ranges[2];
ranges[0] = leveldb::Range("a", "c");
ranges[1] = leveldb::Range("x", "z");
uint64_t sizes[2];
leveldb::Status s = db-&gt;GetApproximateSizes(ranges, 2, sizes);
</pre>
The preceding call will set <code>sizes[0]</code> to the approximate number of
bytes of file system space used by the key range <code>[a..c)</code> and
<code>sizes[1]</code> to the approximate number of bytes used by the key range
<code>[x..z)</code>.
<p>
<h1>Environment</h1>
<p>
All file operations (and other operating system calls) issued by the
<code>leveldb</code> implementation are routed through a <code>leveldb::Env</code> object.
Sophisticated clients may wish to provide their own <code>Env</code>
implementation to get better control. For example, an application may
introduce artificial delays in the file IO paths to limit the impact
of <code>leveldb</code> on other activities in the system.
<p>
<pre>
class SlowEnv : public leveldb::Env {
.. implementation of the Env interface ...
};
SlowEnv env;
leveldb::Options options;
options.env = &amp;env;
Status s = leveldb::DB::Open(options, ...);
</pre>
<h1>Porting</h1>
<p>
<code>leveldb</code> may be ported to a new platform by providing platform
specific implementations of the types/methods/functions exported by
<code>leveldb/port/port.h</code>. See <code>leveldb/port/port_example.h</code> for more
details.
<p>
In addition, the new platform may need a new default <code>leveldb::Env</code>
implementation. See <code>leveldb/util/env_posix.h</code> for an example.
<h1>Other Information</h1>
<p>
Details about the <code>leveldb</code> implementation may be found in
the following documents:
<ul>
<li> <a href="impl.html">Implementation notes</a>
<li> <a href="table_format.txt">Format of an immutable Table file</a>
<li> <a href="log_format.txt">Format of a log file</a>
</ul>
</body>
</html>

View File

@ -1,524 +0,0 @@
leveldb
=======
_Jeff Dean, Sanjay Ghemawat_
The leveldb library provides a persistent key value store. Keys and values are
arbitrary byte arrays. The keys are ordered within the key value store
according to a user-specified comparator function.
## Opening A Database
A leveldb database has a name which corresponds to a file system directory. All
of the contents of database are stored in this directory. The following example
shows how to open a database, creating it if necessary:
```c++
#include <cassert>
#include "leveldb/db.h"
leveldb::DB* db;
leveldb::Options options;
options.create_if_missing = true;
leveldb::Status status = leveldb::DB::Open(options, "/tmp/testdb", &db);
assert(status.ok());
...
```
If you want to raise an error if the database already exists, add the following
line before the `leveldb::DB::Open` call:
```c++
options.error_if_exists = true;
```
## Status
You may have noticed the `leveldb::Status` type above. Values of this type are
returned by most functions in leveldb that may encounter an error. You can check
if such a result is ok, and also print an associated error message:
```c++
leveldb::Status s = ...;
if (!s.ok()) cerr << s.ToString() << endl;
```
## Closing A Database
When you are done with a database, just delete the database object. Example:
```c++
... open the db as described above ...
... do something with db ...
delete db;
```
## Reads And Writes
The database provides Put, Delete, and Get methods to modify/query the database.
For example, the following code moves the value stored under key1 to key2.
```c++
std::string value;
leveldb::Status s = db->Get(leveldb::ReadOptions(), key1, &value);
if (s.ok()) s = db->Put(leveldb::WriteOptions(), key2, value);
if (s.ok()) s = db->Delete(leveldb::WriteOptions(), key1);
```
## Atomic Updates
Note that if the process dies after the Put of key2 but before the delete of
key1, the same value may be left stored under multiple keys. Such problems can
be avoided by using the `WriteBatch` class to atomically apply a set of updates:
```c++
#include "leveldb/write_batch.h"
...
std::string value;
leveldb::Status s = db->Get(leveldb::ReadOptions(), key1, &value);
if (s.ok()) {
leveldb::WriteBatch batch;
batch.Delete(key1);
batch.Put(key2, value);
s = db->Write(leveldb::WriteOptions(), &batch);
}
```
The `WriteBatch` holds a sequence of edits to be made to the database, and these
edits within the batch are applied in order. Note that we called Delete before
Put so that if key1 is identical to key2, we do not end up erroneously dropping
the value entirely.
Apart from its atomicity benefits, `WriteBatch` may also be used to speed up
bulk updates by placing lots of individual mutations into the same batch.
## Synchronous Writes
By default, each write to leveldb is asynchronous: it returns after pushing the
write from the process into the operating system. The transfer from operating
system memory to the underlying persistent storage happens asynchronously. The
sync flag can be turned on for a particular write to make the write operation
not return until the data being written has been pushed all the way to
persistent storage. (On Posix systems, this is implemented by calling either
`fsync(...)` or `fdatasync(...)` or `msync(..., MS_SYNC)` before the write
operation returns.)
```c++
leveldb::WriteOptions write_options;
write_options.sync = true;
db->Put(write_options, ...);
```
Asynchronous writes are often more than a thousand times as fast as synchronous
writes. The downside of asynchronous writes is that a crash of the machine may
cause the last few updates to be lost. Note that a crash of just the writing
process (i.e., not a reboot) will not cause any loss since even when sync is
false, an update is pushed from the process memory into the operating system
before it is considered done.
Asynchronous writes can often be used safely. For example, when loading a large
amount of data into the database you can handle lost updates by restarting the
bulk load after a crash. A hybrid scheme is also possible where every Nth write
is synchronous, and in the event of a crash, the bulk load is restarted just
after the last synchronous write finished by the previous run. (The synchronous
write can update a marker that describes where to restart on a crash.)
`WriteBatch` provides an alternative to asynchronous writes. Multiple updates
may be placed in the same WriteBatch and applied together using a synchronous
write (i.e., `write_options.sync` is set to true). The extra cost of the
synchronous write will be amortized across all of the writes in the batch.
## Concurrency
A database may only be opened by one process at a time. The leveldb
implementation acquires a lock from the operating system to prevent misuse.
Within a single process, the same `leveldb::DB` object may be safely shared by
multiple concurrent threads. I.e., different threads may write into or fetch
iterators or call Get on the same database without any external synchronization
(the leveldb implementation will automatically do the required synchronization).
However other objects (like Iterator and `WriteBatch`) may require external
synchronization. If two threads share such an object, they must protect access
to it using their own locking protocol. More details are available in the public
header files.
## Iteration
The following example demonstrates how to print all key,value pairs in a
database.
```c++
leveldb::Iterator* it = db->NewIterator(leveldb::ReadOptions());
for (it->SeekToFirst(); it->Valid(); it->Next()) {
cout << it->key().ToString() << ": " << it->value().ToString() << endl;
}
assert(it->status().ok()); // Check for any errors found during the scan
delete it;
```
The following variation shows how to process just the keys in the range
[start,limit):
```c++
for (it->Seek(start);
it->Valid() && it->key().ToString() < limit;
it->Next()) {
...
}
```
You can also process entries in reverse order. (Caveat: reverse iteration may be
somewhat slower than forward iteration.)
```c++
for (it->SeekToLast(); it->Valid(); it->Prev()) {
...
}
```
## Snapshots
Snapshots provide consistent read-only views over the entire state of the
key-value store. `ReadOptions::snapshot` may be non-NULL to indicate that a
read should operate on a particular version of the DB state. If
`ReadOptions::snapshot` is NULL, the read will operate on an implicit snapshot
of the current state.
Snapshots are created by the `DB::GetSnapshot()` method:
```c++
leveldb::ReadOptions options;
options.snapshot = db->GetSnapshot();
... apply some updates to db ...
leveldb::Iterator* iter = db->NewIterator(options);
... read using iter to view the state when the snapshot was created ...
delete iter;
db->ReleaseSnapshot(options.snapshot);
```
Note that when a snapshot is no longer needed, it should be released using the
`DB::ReleaseSnapshot` interface. This allows the implementation to get rid of
state that was being maintained just to support reading as of that snapshot.
## Slice
The return value of the `it->key()` and `it->value()` calls above are instances
of the `leveldb::Slice` type. Slice is a simple structure that contains a length
and a pointer to an external byte array. Returning a Slice is a cheaper
alternative to returning a `std::string` since we do not need to copy
potentially large keys and values. In addition, leveldb methods do not return
null-terminated C-style strings since leveldb keys and values are allowed to
contain `'\0'` bytes.
C++ strings and null-terminated C-style strings can be easily converted to a
Slice:
```c++
leveldb::Slice s1 = "hello";
std::string str("world");
leveldb::Slice s2 = str;
```
A Slice can be easily converted back to a C++ string:
```c++
std::string str = s1.ToString();
assert(str == std::string("hello"));
```
Be careful when using Slices since it is up to the caller to ensure that the
external byte array into which the Slice points remains live while the Slice is
in use. For example, the following is buggy:
```c++
leveldb::Slice slice;
if (...) {
std::string str = ...;
slice = str;
}
Use(slice);
```
When the if statement goes out of scope, str will be destroyed and the backing
storage for slice will disappear.
## Comparators
The preceding examples used the default ordering function for key, which orders
bytes lexicographically. You can however supply a custom comparator when opening
a database. For example, suppose each database key consists of two numbers and
we should sort by the first number, breaking ties by the second number. First,
define a proper subclass of `leveldb::Comparator` that expresses these rules:
```c++
class TwoPartComparator : public leveldb::Comparator {
public:
// Three-way comparison function:
// if a < b: negative result
// if a > b: positive result
// else: zero result
int Compare(const leveldb::Slice& a, const leveldb::Slice& b) const {
int a1, a2, b1, b2;
ParseKey(a, &a1, &a2);
ParseKey(b, &b1, &b2);
if (a1 < b1) return -1;
if (a1 > b1) return +1;
if (a2 < b2) return -1;
if (a2 > b2) return +1;
return 0;
}
// Ignore the following methods for now:
const char* Name() const { return "TwoPartComparator"; }
void FindShortestSeparator(std::string*, const leveldb::Slice&) const {}
void FindShortSuccessor(std::string*) const {}
};
```
Now create a database using this custom comparator:
```c++
TwoPartComparator cmp;
leveldb::DB* db;
leveldb::Options options;
options.create_if_missing = true;
options.comparator = &cmp;
leveldb::Status status = leveldb::DB::Open(options, "/tmp/testdb", &db);
...
```
### Backwards compatibility
The result of the comparator's Name method is attached to the database when it
is created, and is checked on every subsequent database open. If the name
changes, the `leveldb::DB::Open` call will fail. Therefore, change the name if
and only if the new key format and comparison function are incompatible with
existing databases, and it is ok to discard the contents of all existing
databases.
You can however still gradually evolve your key format over time with a little
bit of pre-planning. For example, you could store a version number at the end of
each key (one byte should suffice for most uses). When you wish to switch to a
new key format (e.g., adding an optional third part to the keys processed by
`TwoPartComparator`), (a) keep the same comparator name (b) increment the
version number for new keys (c) change the comparator function so it uses the
version numbers found in the keys to decide how to interpret them.
## Performance
Performance can be tuned by changing the default values of the types defined in
`include/options.h`.
### Block size
leveldb groups adjacent keys together into the same block and such a block is
the unit of transfer to and from persistent storage. The default block size is
approximately 4096 uncompressed bytes. Applications that mostly do bulk scans
over the contents of the database may wish to increase this size. Applications
that do a lot of point reads of small values may wish to switch to a smaller
block size if performance measurements indicate an improvement. There isn't much
benefit in using blocks smaller than one kilobyte, or larger than a few
megabytes. Also note that compression will be more effective with larger block
sizes.
### Compression
Each block is individually compressed before being written to persistent
storage. Compression is on by default since the default compression method is
very fast, and is automatically disabled for uncompressible data. In rare cases,
applications may want to disable compression entirely, but should only do so if
benchmarks show a performance improvement:
```c++
leveldb::Options options;
options.compression = leveldb::kNoCompression;
... leveldb::DB::Open(options, name, ...) ....
```
### Cache
The contents of the database are stored in a set of files in the filesystem and
each file stores a sequence of compressed blocks. If options.block_cache is
non-NULL, it is used to cache frequently used uncompressed block contents.
```c++
#include "leveldb/cache.h"
leveldb::Options options;
options.block_cache = leveldb::NewLRUCache(100 * 1048576); // 100MB cache
leveldb::DB* db;
leveldb::DB::Open(options, name, &db);
... use the db ...
delete db
delete options.block_cache;
```
Note that the cache holds uncompressed data, and therefore it should be sized
according to application level data sizes, without any reduction from
compression. (Caching of compressed blocks is left to the operating system
buffer cache, or any custom Env implementation provided by the client.)
When performing a bulk read, the application may wish to disable caching so that
the data processed by the bulk read does not end up displacing most of the
cached contents. A per-iterator option can be used to achieve this:
```c++
leveldb::ReadOptions options;
options.fill_cache = false;
leveldb::Iterator* it = db->NewIterator(options);
for (it->SeekToFirst(); it->Valid(); it->Next()) {
...
}
delete it;
```
### Key Layout
Note that the unit of disk transfer and caching is a block. Adjacent keys
(according to the database sort order) will usually be placed in the same block.
Therefore the application can improve its performance by placing keys that are
accessed together near each other and placing infrequently used keys in a
separate region of the key space.
For example, suppose we are implementing a simple file system on top of leveldb.
The types of entries we might wish to store are:
filename -> permission-bits, length, list of file_block_ids
file_block_id -> data
We might want to prefix filename keys with one letter (say '/') and the
`file_block_id` keys with a different letter (say '0') so that scans over just
the metadata do not force us to fetch and cache bulky file contents.
### Filters
Because of the way leveldb data is organized on disk, a single `Get()` call may
involve multiple reads from disk. The optional FilterPolicy mechanism can be
used to reduce the number of disk reads substantially.
```c++
leveldb::Options options;
options.filter_policy = NewBloomFilterPolicy(10);
leveldb::DB* db;
leveldb::DB::Open(options, "/tmp/testdb", &db);
... use the database ...
delete db;
delete options.filter_policy;
```
The preceding code associates a Bloom filter based filtering policy with the
database. Bloom filter based filtering relies on keeping some number of bits of
data in memory per key (in this case 10 bits per key since that is the argument
we passed to `NewBloomFilterPolicy`). This filter will reduce the number of
unnecessary disk reads needed for Get() calls by a factor of approximately
a 100. Increasing the bits per key will lead to a larger reduction at the cost
of more memory usage. We recommend that applications whose working set does not
fit in memory and that do a lot of random reads set a filter policy.
If you are using a custom comparator, you should ensure that the filter policy
you are using is compatible with your comparator. For example, consider a
comparator that ignores trailing spaces when comparing keys.
`NewBloomFilterPolicy` must not be used with such a comparator. Instead, the
application should provide a custom filter policy that also ignores trailing
spaces. For example:
```c++
class CustomFilterPolicy : public leveldb::FilterPolicy {
private:
leveldb::FilterPolicy* builtin_policy_;
public:
CustomFilterPolicy() : builtin_policy_(leveldb::NewBloomFilterPolicy(10)) {}
~CustomFilterPolicy() { delete builtin_policy_; }
const char* Name() const { return "IgnoreTrailingSpacesFilter"; }
void CreateFilter(const leveldb::Slice* keys, int n, std::string* dst) const {
// Use builtin bloom filter code after removing trailing spaces
std::vector<leveldb::Slice> trimmed(n);
for (int i = 0; i < n; i++) {
trimmed[i] = RemoveTrailingSpaces(keys[i]);
}
builtin_policy_->CreateFilter(trimmed.data(), n, dst);
}
};
```
Advanced applications may provide a filter policy that does not use a bloom
filter but uses some other mechanism for summarizing a set of keys. See
`leveldb/filter_policy.h` for detail.
## Checksums
leveldb associates checksums with all data it stores in the file system. There
are two separate controls provided over how aggressively these checksums are
verified:
`ReadOptions::verify_checksums` may be set to true to force checksum
verification of all data that is read from the file system on behalf of a
particular read. By default, no such verification is done.
`Options::paranoid_checks` may be set to true before opening a database to make
the database implementation raise an error as soon as it detects an internal
corruption. Depending on which portion of the database has been corrupted, the
error may be raised when the database is opened, or later by another database
operation. By default, paranoid checking is off so that the database can be used
even if parts of its persistent storage have been corrupted.
If a database is corrupted (perhaps it cannot be opened when paranoid checking
is turned on), the `leveldb::RepairDB` function may be used to recover as much
of the data as possible
## Approximate Sizes
The `GetApproximateSizes` method can used to get the approximate number of bytes
of file system space used by one or more key ranges.
```c++
leveldb::Range ranges[2];
ranges[0] = leveldb::Range("a", "c");
ranges[1] = leveldb::Range("x", "z");
uint64_t sizes[2];
db->GetApproximateSizes(ranges, 2, sizes);
```
The preceding call will set `sizes[0]` to the approximate number of bytes of
file system space used by the key range `[a..c)` and `sizes[1]` to the
approximate number of bytes used by the key range `[x..z)`.
## Environment
All file operations (and other operating system calls) issued by the leveldb
implementation are routed through a `leveldb::Env` object. Sophisticated clients
may wish to provide their own Env implementation to get better control.
For example, an application may introduce artificial delays in the file IO
paths to limit the impact of leveldb on other activities in the system.
```c++
class SlowEnv : public leveldb::Env {
... implementation of the Env interface ...
};
SlowEnv env;
leveldb::Options options;
options.env = &env;
Status s = leveldb::DB::Open(options, ...);
```
## Porting
leveldb may be ported to a new platform by providing platform specific
implementations of the types/methods/functions exported by
`leveldb/port/port.h`. See `leveldb/port/port_example.h` for more details.
In addition, the new platform may need a new default `leveldb::Env`
implementation. See `leveldb/util/env_posix.h` for an example.
## Other Information
Details about the leveldb implementation may be found in the following
documents:
1. [Implementation notes](impl.md)
2. [Format of an immutable Table file](table_format.md)
3. [Format of a log file](log_format.md)

View File

@ -1,75 +0,0 @@
leveldb Log format
==================
The log file contents are a sequence of 32KB blocks. The only exception is that
the tail of the file may contain a partial block.
Each block consists of a sequence of records:
block := record* trailer?
record :=
checksum: uint32 // crc32c of type and data[] ; little-endian
length: uint16 // little-endian
type: uint8 // One of FULL, FIRST, MIDDLE, LAST
data: uint8[length]
A record never starts within the last six bytes of a block (since it won't fit).
Any leftover bytes here form the trailer, which must consist entirely of zero
bytes and must be skipped by readers.
Aside: if exactly seven bytes are left in the current block, and a new non-zero
length record is added, the writer must emit a FIRST record (which contains zero
bytes of user data) to fill up the trailing seven bytes of the block and then
emit all of the user data in subsequent blocks.
More types may be added in the future. Some Readers may skip record types they
do not understand, others may report that some data was skipped.
FULL == 1
FIRST == 2
MIDDLE == 3
LAST == 4
The FULL record contains the contents of an entire user record.
FIRST, MIDDLE, LAST are types used for user records that have been split into
multiple fragments (typically because of block boundaries). FIRST is the type
of the first fragment of a user record, LAST is the type of the last fragment of
a user record, and MIDDLE is the type of all interior fragments of a user
record.
Example: consider a sequence of user records:
A: length 1000
B: length 97270
C: length 8000
**A** will be stored as a FULL record in the first block.
**B** will be split into three fragments: first fragment occupies the rest of
the first block, second fragment occupies the entirety of the second block, and
the third fragment occupies a prefix of the third block. This will leave six
bytes free in the third block, which will be left empty as the trailer.
**C** will be stored as a FULL record in the fourth block.
----
## Some benefits over the recordio format:
1. We do not need any heuristics for resyncing - just go to next block boundary
and scan. If there is a corruption, skip to the next block. As a
side-benefit, we do not get confused when part of the contents of one log
file are embedded as a record inside another log file.
2. Splitting at approximate boundaries (e.g., for mapreduce) is simple: find the
next block boundary and skip records until we hit a FULL or FIRST record.
3. We do not need extra buffering for large records.
## Some downsides compared to recordio format:
1. No packing of tiny records. This could be fixed by adding a new record type,
so it is a shortcoming of the current implementation, not necessarily the
format.
2. No compression. Again, this could be fixed by adding new record types.

75
doc/log_format.txt Normal file
View File

@ -0,0 +1,75 @@
The log file contents are a sequence of 32KB blocks. The only
exception is that the tail of the file may contain a partial block.
Each block consists of a sequence of records:
block := record* trailer?
record :=
checksum: uint32 // crc32c of type and data[] ; little-endian
length: uint16 // little-endian
type: uint8 // One of FULL, FIRST, MIDDLE, LAST
data: uint8[length]
A record never starts within the last six bytes of a block (since it
won't fit). Any leftover bytes here form the trailer, which must
consist entirely of zero bytes and must be skipped by readers.
Aside: if exactly seven bytes are left in the current block, and a new
non-zero length record is added, the writer must emit a FIRST record
(which contains zero bytes of user data) to fill up the trailing seven
bytes of the block and then emit all of the user data in subsequent
blocks.
More types may be added in the future. Some Readers may skip record
types they do not understand, others may report that some data was
skipped.
FULL == 1
FIRST == 2
MIDDLE == 3
LAST == 4
The FULL record contains the contents of an entire user record.
FIRST, MIDDLE, LAST are types used for user records that have been
split into multiple fragments (typically because of block boundaries).
FIRST is the type of the first fragment of a user record, LAST is the
type of the last fragment of a user record, and MIDDLE is the type of
all interior fragments of a user record.
Example: consider a sequence of user records:
A: length 1000
B: length 97270
C: length 8000
A will be stored as a FULL record in the first block.
B will be split into three fragments: first fragment occupies the rest
of the first block, second fragment occupies the entirety of the
second block, and the third fragment occupies a prefix of the third
block. This will leave six bytes free in the third block, which will
be left empty as the trailer.
C will be stored as a FULL record in the fourth block.
===================
Some benefits over the recordio format:
(1) We do not need any heuristics for resyncing - just go to next
block boundary and scan. If there is a corruption, skip to the next
block. As a side-benefit, we do not get confused when part of the
contents of one log file are embedded as a record inside another log
file.
(2) Splitting at approximate boundaries (e.g., for mapreduce) is
simple: find the next block boundary and skip records until we
hit a FULL or FIRST record.
(3) We do not need extra buffering for large records.
Some downsides compared to recordio format:
(1) No packing of tiny records. This could be fixed by adding a new
record type, so it is a shortcoming of the current implementation,
not necessarily the format.
(2) No compression. Again, this could be fixed by adding new record types.

View File

@ -1,107 +0,0 @@
leveldb File format
===================
<beginning_of_file>
[data block 1]
[data block 2]
...
[data block N]
[meta block 1]
...
[meta block K]
[metaindex block]
[index block]
[Footer] (fixed size; starts at file_size - sizeof(Footer))
<end_of_file>
The file contains internal pointers. Each such pointer is called
a BlockHandle and contains the following information:
offset: varint64
size: varint64
See [varints](https://developers.google.com/protocol-buffers/docs/encoding#varints)
for an explanation of varint64 format.
1. The sequence of key/value pairs in the file are stored in sorted
order and partitioned into a sequence of data blocks. These blocks
come one after another at the beginning of the file. Each data block
is formatted according to the code in `block_builder.cc`, and then
optionally compressed.
2. After the data blocks we store a bunch of meta blocks. The
supported meta block types are described below. More meta block types
may be added in the future. Each meta block is again formatted using
`block_builder.cc` and then optionally compressed.
3. A "metaindex" block. It contains one entry for every other meta
block where the key is the name of the meta block and the value is a
BlockHandle pointing to that meta block.
4. An "index" block. This block contains one entry per data block,
where the key is a string >= last key in that data block and before
the first key in the successive data block. The value is the
BlockHandle for the data block.
5. At the very end of the file is a fixed length footer that contains
the BlockHandle of the metaindex and index blocks as well as a magic number.
metaindex_handle: char[p]; // Block handle for metaindex
index_handle: char[q]; // Block handle for index
padding: char[40-p-q];// zeroed bytes to make fixed length
// (40==2*BlockHandle::kMaxEncodedLength)
magic: fixed64; // == 0xdb4775248b80fb57 (little-endian)
## "filter" Meta Block
If a `FilterPolicy` was specified when the database was opened, a
filter block is stored in each table. The "metaindex" block contains
an entry that maps from `filter.<N>` to the BlockHandle for the filter
block where `<N>` is the string returned by the filter policy's
`Name()` method.
The filter block stores a sequence of filters, where filter i contains
the output of `FilterPolicy::CreateFilter()` on all keys that are stored
in a block whose file offset falls within the range
[ i*base ... (i+1)*base-1 ]
Currently, "base" is 2KB. So for example, if blocks X and Y start in
the range `[ 0KB .. 2KB-1 ]`, all of the keys in X and Y will be
converted to a filter by calling `FilterPolicy::CreateFilter()`, and the
resulting filter will be stored as the first filter in the filter
block.
The filter block is formatted as follows:
[filter 0]
[filter 1]
[filter 2]
...
[filter N-1]
[offset of filter 0] : 4 bytes
[offset of filter 1] : 4 bytes
[offset of filter 2] : 4 bytes
...
[offset of filter N-1] : 4 bytes
[offset of beginning of offset array] : 4 bytes
lg(base) : 1 byte
The offset array at the end of the filter block allows efficient
mapping from a data block offset to the corresponding filter.
## "stats" Meta Block
This meta block contains a bunch of stats. The key is the name
of the statistic. The value contains the statistic.
TODO(postrelease): record following stats.
data size
index size
key size (uncompressed)
value size (uncompressed)
number of entries
number of data blocks

104
doc/table_format.txt Normal file
View File

@ -0,0 +1,104 @@
File format
===========
<beginning_of_file>
[data block 1]
[data block 2]
...
[data block N]
[meta block 1]
...
[meta block K]
[metaindex block]
[index block]
[Footer] (fixed size; starts at file_size - sizeof(Footer))
<end_of_file>
The file contains internal pointers. Each such pointer is called
a BlockHandle and contains the following information:
offset: varint64
size: varint64
See https://developers.google.com/protocol-buffers/docs/encoding#varints
for an explanation of varint64 format.
(1) The sequence of key/value pairs in the file are stored in sorted
order and partitioned into a sequence of data blocks. These blocks
come one after another at the beginning of the file. Each data block
is formatted according to the code in block_builder.cc, and then
optionally compressed.
(2) After the data blocks we store a bunch of meta blocks. The
supported meta block types are described below. More meta block types
may be added in the future. Each meta block is again formatted using
block_builder.cc and then optionally compressed.
(3) A "metaindex" block. It contains one entry for every other meta
block where the key is the name of the meta block and the value is a
BlockHandle pointing to that meta block.
(4) An "index" block. This block contains one entry per data block,
where the key is a string >= last key in that data block and before
the first key in the successive data block. The value is the
BlockHandle for the data block.
(6) At the very end of the file is a fixed length footer that contains
the BlockHandle of the metaindex and index blocks as well as a magic number.
metaindex_handle: char[p]; // Block handle for metaindex
index_handle: char[q]; // Block handle for index
padding: char[40-p-q]; // zeroed bytes to make fixed length
// (40==2*BlockHandle::kMaxEncodedLength)
magic: fixed64; // == 0xdb4775248b80fb57 (little-endian)
"filter" Meta Block
-------------------
If a "FilterPolicy" was specified when the database was opened, a
filter block is stored in each table. The "metaindex" block contains
an entry that maps from "filter.<N>" to the BlockHandle for the filter
block where "<N>" is the string returned by the filter policy's
"Name()" method.
The filter block stores a sequence of filters, where filter i contains
the output of FilterPolicy::CreateFilter() on all keys that are stored
in a block whose file offset falls within the range
[ i*base ... (i+1)*base-1 ]
Currently, "base" is 2KB. So for example, if blocks X and Y start in
the range [ 0KB .. 2KB-1 ], all of the keys in X and Y will be
converted to a filter by calling FilterPolicy::CreateFilter(), and the
resulting filter will be stored as the first filter in the filter
block.
The filter block is formatted as follows:
[filter 0]
[filter 1]
[filter 2]
...
[filter N-1]
[offset of filter 0] : 4 bytes
[offset of filter 1] : 4 bytes
[offset of filter 2] : 4 bytes
...
[offset of filter N-1] : 4 bytes
[offset of beginning of offset array] : 4 bytes
lg(base) : 1 byte
The offset array at the end of the filter block allows efficient
mapping from a data block offset to the corresponding filter.
"stats" Meta Block
------------------
This meta block contains a bunch of stats. The key is the name
of the statistic. The value contains the statistic.
TODO(postrelease): record following stats.
data size
index size
key size (uncompressed)
value size (uncompressed)
number of entries
number of data blocks

View File

@ -4,17 +4,14 @@
#include "helpers/memenv/memenv.h" #include "helpers/memenv/memenv.h"
#include <cstring>
#include <limits>
#include <map>
#include <string>
#include <vector>
#include "leveldb/env.h" #include "leveldb/env.h"
#include "leveldb/status.h" #include "leveldb/status.h"
#include "port/port.h" #include "port/port.h"
#include "port/thread_annotations.h"
#include "util/mutexlock.h" #include "util/mutexlock.h"
#include <map>
#include <string.h>
#include <string>
#include <vector>
namespace leveldb { namespace leveldb {
@ -26,10 +23,6 @@ class FileState {
// and the caller must call Ref() at least once. // and the caller must call Ref() at least once.
FileState() : refs_(0), size_(0) {} FileState() : refs_(0), size_(0) {}
// No copying allowed.
FileState(const FileState&) = delete;
FileState& operator=(const FileState&) = delete;
// Increase the reference count. // Increase the reference count.
void Ref() { void Ref() {
MutexLock lock(&refs_mutex_); MutexLock lock(&refs_mutex_);
@ -54,22 +47,9 @@ class FileState {
} }
} }
uint64_t Size() const { uint64_t Size() const { return size_; }
MutexLock lock(&blocks_mutex_);
return size_;
}
void Truncate() {
MutexLock lock(&blocks_mutex_);
for (char*& block : blocks_) {
delete[] block;
}
blocks_.clear();
size_ = 0;
}
Status Read(uint64_t offset, size_t n, Slice* result, char* scratch) const { Status Read(uint64_t offset, size_t n, Slice* result, char* scratch) const {
MutexLock lock(&blocks_mutex_);
if (offset > size_) { if (offset > size_) {
return Status::IOError("Offset greater than file size."); return Status::IOError("Offset greater than file size.");
} }
@ -82,9 +62,16 @@ class FileState {
return Status::OK(); return Status::OK();
} }
assert(offset / kBlockSize <= std::numeric_limits<size_t>::max()); assert(offset / kBlockSize <= SIZE_MAX);
size_t block = static_cast<size_t>(offset / kBlockSize); size_t block = static_cast<size_t>(offset / kBlockSize);
size_t block_offset = offset % kBlockSize; size_t block_offset = offset % kBlockSize;
if (n <= kBlockSize - block_offset) {
// The requested bytes are all in the first block.
*result = Slice(blocks_[block] + block_offset, n);
return Status::OK();
}
size_t bytes_to_copy = n; size_t bytes_to_copy = n;
char* dst = scratch; char* dst = scratch;
@ -93,7 +80,7 @@ class FileState {
if (avail > bytes_to_copy) { if (avail > bytes_to_copy) {
avail = bytes_to_copy; avail = bytes_to_copy;
} }
std::memcpy(dst, blocks_[block] + block_offset, avail); memcpy(dst, blocks_[block] + block_offset, avail);
bytes_to_copy -= avail; bytes_to_copy -= avail;
dst += avail; dst += avail;
@ -109,7 +96,6 @@ class FileState {
const char* src = data.data(); const char* src = data.data();
size_t src_len = data.size(); size_t src_len = data.size();
MutexLock lock(&blocks_mutex_);
while (src_len > 0) { while (src_len > 0) {
size_t avail; size_t avail;
size_t offset = size_ % kBlockSize; size_t offset = size_ % kBlockSize;
@ -126,7 +112,7 @@ class FileState {
if (avail > src_len) { if (avail > src_len) {
avail = src_len; avail = src_len;
} }
std::memcpy(blocks_.back() + offset, src, avail); memcpy(blocks_.back() + offset, src, avail);
src_len -= avail; src_len -= avail;
src += avail; src += avail;
size_ += avail; size_ += avail;
@ -136,17 +122,28 @@ class FileState {
} }
private: private:
enum { kBlockSize = 8 * 1024 };
// Private since only Unref() should be used to delete it. // Private since only Unref() should be used to delete it.
~FileState() { Truncate(); } ~FileState() {
for (std::vector<char*>::iterator i = blocks_.begin(); i != blocks_.end();
++i) {
delete [] *i;
}
}
// No copying allowed.
FileState(const FileState&);
void operator=(const FileState&);
port::Mutex refs_mutex_; port::Mutex refs_mutex_;
int refs_ GUARDED_BY(refs_mutex_); int refs_; // Protected by refs_mutex_;
mutable port::Mutex blocks_mutex_; // The following fields are not protected by any mutex. They are only mutable
std::vector<char*> blocks_ GUARDED_BY(blocks_mutex_); // while the file is being written, and concurrent access is not allowed
uint64_t size_ GUARDED_BY(blocks_mutex_); // to writable files.
std::vector<char*> blocks_;
uint64_t size_;
enum { kBlockSize = 8 * 1024 };
}; };
class SequentialFileImpl : public SequentialFile { class SequentialFileImpl : public SequentialFile {
@ -155,9 +152,11 @@ class SequentialFileImpl : public SequentialFile {
file_->Ref(); file_->Ref();
} }
~SequentialFileImpl() override { file_->Unref(); } ~SequentialFileImpl() {
file_->Unref();
}
Status Read(size_t n, Slice* result, char* scratch) override { virtual Status Read(size_t n, Slice* result, char* scratch) {
Status s = file_->Read(pos_, n, result, scratch); Status s = file_->Read(pos_, n, result, scratch);
if (s.ok()) { if (s.ok()) {
pos_ += result->size(); pos_ += result->size();
@ -165,7 +164,7 @@ class SequentialFileImpl : public SequentialFile {
return s; return s;
} }
Status Skip(uint64_t n) override { virtual Status Skip(uint64_t n) {
if (pos_ > file_->Size()) { if (pos_ > file_->Size()) {
return Status::IOError("pos_ > file_->Size()"); return Status::IOError("pos_ > file_->Size()");
} }
@ -184,12 +183,16 @@ class SequentialFileImpl : public SequentialFile {
class RandomAccessFileImpl : public RandomAccessFile { class RandomAccessFileImpl : public RandomAccessFile {
public: public:
explicit RandomAccessFileImpl(FileState* file) : file_(file) { file_->Ref(); } explicit RandomAccessFileImpl(FileState* file) : file_(file) {
file_->Ref();
}
~RandomAccessFileImpl() override { file_->Unref(); } ~RandomAccessFileImpl() {
file_->Unref();
}
Status Read(uint64_t offset, size_t n, Slice* result, virtual Status Read(uint64_t offset, size_t n, Slice* result,
char* scratch) const override { char* scratch) const {
return file_->Read(offset, n, result, scratch); return file_->Read(offset, n, result, scratch);
} }
@ -199,15 +202,21 @@ class RandomAccessFileImpl : public RandomAccessFile {
class WritableFileImpl : public WritableFile { class WritableFileImpl : public WritableFile {
public: public:
WritableFileImpl(FileState* file) : file_(file) { file_->Ref(); } WritableFileImpl(FileState* file) : file_(file) {
file_->Ref();
}
~WritableFileImpl() override { file_->Unref(); } ~WritableFileImpl() {
file_->Unref();
}
Status Append(const Slice& data) override { return file_->Append(data); } virtual Status Append(const Slice& data) {
return file_->Append(data);
}
Status Close() override { return Status::OK(); } virtual Status Close() { return Status::OK(); }
Status Flush() override { return Status::OK(); } virtual Status Flush() { return Status::OK(); }
Status Sync() override { return Status::OK(); } virtual Status Sync() { return Status::OK(); }
private: private:
FileState* file_; FileState* file_;
@ -215,25 +224,25 @@ class WritableFileImpl : public WritableFile {
class NoOpLogger : public Logger { class NoOpLogger : public Logger {
public: public:
void Logv(const char* format, std::va_list ap) override {} virtual void Logv(const char* format, va_list ap) { }
}; };
class InMemoryEnv : public EnvWrapper { class InMemoryEnv : public EnvWrapper {
public: public:
explicit InMemoryEnv(Env* base_env) : EnvWrapper(base_env) {} explicit InMemoryEnv(Env* base_env) : EnvWrapper(base_env) { }
~InMemoryEnv() override { virtual ~InMemoryEnv() {
for (const auto& kvp : file_map_) { for (FileSystem::iterator i = file_map_.begin(); i != file_map_.end(); ++i){
kvp.second->Unref(); i->second->Unref();
} }
} }
// Partial implementation of the Env interface. // Partial implementation of the Env interface.
Status NewSequentialFile(const std::string& fname, virtual Status NewSequentialFile(const std::string& fname,
SequentialFile** result) override { SequentialFile** result) {
MutexLock lock(&mutex_); MutexLock lock(&mutex_);
if (file_map_.find(fname) == file_map_.end()) { if (file_map_.find(fname) == file_map_.end()) {
*result = nullptr; *result = NULL;
return Status::IOError(fname, "File not found"); return Status::IOError(fname, "File not found");
} }
@ -241,11 +250,11 @@ class InMemoryEnv : public EnvWrapper {
return Status::OK(); return Status::OK();
} }
Status NewRandomAccessFile(const std::string& fname, virtual Status NewRandomAccessFile(const std::string& fname,
RandomAccessFile** result) override { RandomAccessFile** result) {
MutexLock lock(&mutex_); MutexLock lock(&mutex_);
if (file_map_.find(fname) == file_map_.end()) { if (file_map_.find(fname) == file_map_.end()) {
*result = nullptr; *result = NULL;
return Status::IOError(fname, "File not found"); return Status::IOError(fname, "File not found");
} }
@ -253,32 +262,27 @@ class InMemoryEnv : public EnvWrapper {
return Status::OK(); return Status::OK();
} }
Status NewWritableFile(const std::string& fname, virtual Status NewWritableFile(const std::string& fname,
WritableFile** result) override { WritableFile** result) {
MutexLock lock(&mutex_); MutexLock lock(&mutex_);
FileSystem::iterator it = file_map_.find(fname); if (file_map_.find(fname) != file_map_.end()) {
DeleteFileInternal(fname);
FileState* file;
if (it == file_map_.end()) {
// File is not currently open.
file = new FileState();
file->Ref();
file_map_[fname] = file;
} else {
file = it->second;
file->Truncate();
} }
FileState* file = new FileState();
file->Ref();
file_map_[fname] = file;
*result = new WritableFileImpl(file); *result = new WritableFileImpl(file);
return Status::OK(); return Status::OK();
} }
Status NewAppendableFile(const std::string& fname, virtual Status NewAppendableFile(const std::string& fname,
WritableFile** result) override { WritableFile** result) {
MutexLock lock(&mutex_); MutexLock lock(&mutex_);
FileState** sptr = &file_map_[fname]; FileState** sptr = &file_map_[fname];
FileState* file = *sptr; FileState* file = *sptr;
if (file == nullptr) { if (file == NULL) {
file = new FileState(); file = new FileState();
file->Ref(); file->Ref();
} }
@ -286,18 +290,18 @@ class InMemoryEnv : public EnvWrapper {
return Status::OK(); return Status::OK();
} }
bool FileExists(const std::string& fname) override { virtual bool FileExists(const std::string& fname) {
MutexLock lock(&mutex_); MutexLock lock(&mutex_);
return file_map_.find(fname) != file_map_.end(); return file_map_.find(fname) != file_map_.end();
} }
Status GetChildren(const std::string& dir, virtual Status GetChildren(const std::string& dir,
std::vector<std::string>* result) override { std::vector<std::string>* result) {
MutexLock lock(&mutex_); MutexLock lock(&mutex_);
result->clear(); result->clear();
for (const auto& kvp : file_map_) { for (FileSystem::iterator i = file_map_.begin(); i != file_map_.end(); ++i){
const std::string& filename = kvp.first; const std::string& filename = i->first;
if (filename.size() >= dir.size() + 1 && filename[dir.size()] == '/' && if (filename.size() >= dir.size() + 1 && filename[dir.size()] == '/' &&
Slice(filename).starts_with(Slice(dir))) { Slice(filename).starts_with(Slice(dir))) {
@ -308,8 +312,7 @@ class InMemoryEnv : public EnvWrapper {
return Status::OK(); return Status::OK();
} }
void RemoveFileInternal(const std::string& fname) void DeleteFileInternal(const std::string& fname) {
EXCLUSIVE_LOCKS_REQUIRED(mutex_) {
if (file_map_.find(fname) == file_map_.end()) { if (file_map_.find(fname) == file_map_.end()) {
return; return;
} }
@ -318,21 +321,25 @@ class InMemoryEnv : public EnvWrapper {
file_map_.erase(fname); file_map_.erase(fname);
} }
Status RemoveFile(const std::string& fname) override { virtual Status DeleteFile(const std::string& fname) {
MutexLock lock(&mutex_); MutexLock lock(&mutex_);
if (file_map_.find(fname) == file_map_.end()) { if (file_map_.find(fname) == file_map_.end()) {
return Status::IOError(fname, "File not found"); return Status::IOError(fname, "File not found");
} }
RemoveFileInternal(fname); DeleteFileInternal(fname);
return Status::OK(); return Status::OK();
} }
Status CreateDir(const std::string& dirname) override { return Status::OK(); } virtual Status CreateDir(const std::string& dirname) {
return Status::OK();
}
Status RemoveDir(const std::string& dirname) override { return Status::OK(); } virtual Status DeleteDir(const std::string& dirname) {
return Status::OK();
}
Status GetFileSize(const std::string& fname, uint64_t* file_size) override { virtual Status GetFileSize(const std::string& fname, uint64_t* file_size) {
MutexLock lock(&mutex_); MutexLock lock(&mutex_);
if (file_map_.find(fname) == file_map_.end()) { if (file_map_.find(fname) == file_map_.end()) {
return Status::IOError(fname, "File not found"); return Status::IOError(fname, "File not found");
@ -342,35 +349,35 @@ class InMemoryEnv : public EnvWrapper {
return Status::OK(); return Status::OK();
} }
Status RenameFile(const std::string& src, virtual Status RenameFile(const std::string& src,
const std::string& target) override { const std::string& target) {
MutexLock lock(&mutex_); MutexLock lock(&mutex_);
if (file_map_.find(src) == file_map_.end()) { if (file_map_.find(src) == file_map_.end()) {
return Status::IOError(src, "File not found"); return Status::IOError(src, "File not found");
} }
RemoveFileInternal(target); DeleteFileInternal(target);
file_map_[target] = file_map_[src]; file_map_[target] = file_map_[src];
file_map_.erase(src); file_map_.erase(src);
return Status::OK(); return Status::OK();
} }
Status LockFile(const std::string& fname, FileLock** lock) override { virtual Status LockFile(const std::string& fname, FileLock** lock) {
*lock = new FileLock; *lock = new FileLock;
return Status::OK(); return Status::OK();
} }
Status UnlockFile(FileLock* lock) override { virtual Status UnlockFile(FileLock* lock) {
delete lock; delete lock;
return Status::OK(); return Status::OK();
} }
Status GetTestDirectory(std::string* path) override { virtual Status GetTestDirectory(std::string* path) {
*path = "/test"; *path = "/test";
return Status::OK(); return Status::OK();
} }
Status NewLogger(const std::string& fname, Logger** result) override { virtual Status NewLogger(const std::string& fname, Logger** result) {
*result = new NoOpLogger; *result = new NoOpLogger;
return Status::OK(); return Status::OK();
} }
@ -378,13 +385,14 @@ class InMemoryEnv : public EnvWrapper {
private: private:
// Map from filenames to FileState objects, representing a simple file system. // Map from filenames to FileState objects, representing a simple file system.
typedef std::map<std::string, FileState*> FileSystem; typedef std::map<std::string, FileState*> FileSystem;
port::Mutex mutex_; port::Mutex mutex_;
FileSystem file_map_ GUARDED_BY(mutex_); FileSystem file_map_; // Protected by mutex_.
}; };
} // namespace } // namespace
Env* NewMemEnv(Env* base_env) { return new InMemoryEnv(base_env); } Env* NewMemEnv(Env* base_env) {
return new InMemoryEnv(base_env);
}
} // namespace leveldb } // namespace leveldb

View File

@ -5,8 +5,6 @@
#ifndef STORAGE_LEVELDB_HELPERS_MEMENV_MEMENV_H_ #ifndef STORAGE_LEVELDB_HELPERS_MEMENV_MEMENV_H_
#define STORAGE_LEVELDB_HELPERS_MEMENV_MEMENV_H_ #define STORAGE_LEVELDB_HELPERS_MEMENV_MEMENV_H_
#include "leveldb/export.h"
namespace leveldb { namespace leveldb {
class Env; class Env;
@ -15,7 +13,7 @@ class Env;
// all non-file-storage tasks to base_env. The caller must delete the result // all non-file-storage tasks to base_env. The caller must delete the result
// when it is no longer needed. // when it is no longer needed.
// *base_env must remain live while the result is in use. // *base_env must remain live while the result is in use.
LEVELDB_EXPORT Env* NewMemEnv(Env* base_env); Env* NewMemEnv(Env* base_env);
} // namespace leveldb } // namespace leveldb

View File

@ -4,74 +4,76 @@
#include "helpers/memenv/memenv.h" #include "helpers/memenv/memenv.h"
#include <string>
#include <vector>
#include "gtest/gtest.h"
#include "db/db_impl.h" #include "db/db_impl.h"
#include "leveldb/db.h" #include "leveldb/db.h"
#include "leveldb/env.h" #include "leveldb/env.h"
#include "util/testutil.h" #include "util/testharness.h"
#include <string>
#include <vector>
namespace leveldb { namespace leveldb {
class MemEnvTest : public testing::Test { class MemEnvTest {
public: public:
MemEnvTest() : env_(NewMemEnv(Env::Default())) {}
~MemEnvTest() { delete env_; }
Env* env_; Env* env_;
MemEnvTest()
: env_(NewMemEnv(Env::Default())) {
}
~MemEnvTest() {
delete env_;
}
}; };
TEST_F(MemEnvTest, Basics) { TEST(MemEnvTest, Basics) {
uint64_t file_size; uint64_t file_size;
WritableFile* writable_file; WritableFile* writable_file;
std::vector<std::string> children; std::vector<std::string> children;
ASSERT_LEVELDB_OK(env_->CreateDir("/dir")); ASSERT_OK(env_->CreateDir("/dir"));
// Check that the directory is empty. // Check that the directory is empty.
ASSERT_TRUE(!env_->FileExists("/dir/non_existent")); ASSERT_TRUE(!env_->FileExists("/dir/non_existent"));
ASSERT_TRUE(!env_->GetFileSize("/dir/non_existent", &file_size).ok()); ASSERT_TRUE(!env_->GetFileSize("/dir/non_existent", &file_size).ok());
ASSERT_LEVELDB_OK(env_->GetChildren("/dir", &children)); ASSERT_OK(env_->GetChildren("/dir", &children));
ASSERT_EQ(0, children.size()); ASSERT_EQ(0, children.size());
// Create a file. // Create a file.
ASSERT_LEVELDB_OK(env_->NewWritableFile("/dir/f", &writable_file)); ASSERT_OK(env_->NewWritableFile("/dir/f", &writable_file));
ASSERT_LEVELDB_OK(env_->GetFileSize("/dir/f", &file_size)); ASSERT_OK(env_->GetFileSize("/dir/f", &file_size));
ASSERT_EQ(0, file_size); ASSERT_EQ(0, file_size);
delete writable_file; delete writable_file;
// Check that the file exists. // Check that the file exists.
ASSERT_TRUE(env_->FileExists("/dir/f")); ASSERT_TRUE(env_->FileExists("/dir/f"));
ASSERT_LEVELDB_OK(env_->GetFileSize("/dir/f", &file_size)); ASSERT_OK(env_->GetFileSize("/dir/f", &file_size));
ASSERT_EQ(0, file_size); ASSERT_EQ(0, file_size);
ASSERT_LEVELDB_OK(env_->GetChildren("/dir", &children)); ASSERT_OK(env_->GetChildren("/dir", &children));
ASSERT_EQ(1, children.size()); ASSERT_EQ(1, children.size());
ASSERT_EQ("f", children[0]); ASSERT_EQ("f", children[0]);
// Write to the file. // Write to the file.
ASSERT_LEVELDB_OK(env_->NewWritableFile("/dir/f", &writable_file)); ASSERT_OK(env_->NewWritableFile("/dir/f", &writable_file));
ASSERT_LEVELDB_OK(writable_file->Append("abc")); ASSERT_OK(writable_file->Append("abc"));
delete writable_file; delete writable_file;
// Check that append works. // Check that append works.
ASSERT_LEVELDB_OK(env_->NewAppendableFile("/dir/f", &writable_file)); ASSERT_OK(env_->NewAppendableFile("/dir/f", &writable_file));
ASSERT_LEVELDB_OK(env_->GetFileSize("/dir/f", &file_size)); ASSERT_OK(env_->GetFileSize("/dir/f", &file_size));
ASSERT_EQ(3, file_size); ASSERT_EQ(3, file_size);
ASSERT_LEVELDB_OK(writable_file->Append("hello")); ASSERT_OK(writable_file->Append("hello"));
delete writable_file; delete writable_file;
// Check for expected size. // Check for expected size.
ASSERT_LEVELDB_OK(env_->GetFileSize("/dir/f", &file_size)); ASSERT_OK(env_->GetFileSize("/dir/f", &file_size));
ASSERT_EQ(8, file_size); ASSERT_EQ(8, file_size);
// Check that renaming works. // Check that renaming works.
ASSERT_TRUE(!env_->RenameFile("/dir/non_existent", "/dir/g").ok()); ASSERT_TRUE(!env_->RenameFile("/dir/non_existent", "/dir/g").ok());
ASSERT_LEVELDB_OK(env_->RenameFile("/dir/f", "/dir/g")); ASSERT_OK(env_->RenameFile("/dir/f", "/dir/g"));
ASSERT_TRUE(!env_->FileExists("/dir/f")); ASSERT_TRUE(!env_->FileExists("/dir/f"));
ASSERT_TRUE(env_->FileExists("/dir/g")); ASSERT_TRUE(env_->FileExists("/dir/g"));
ASSERT_LEVELDB_OK(env_->GetFileSize("/dir/g", &file_size)); ASSERT_OK(env_->GetFileSize("/dir/g", &file_size));
ASSERT_EQ(8, file_size); ASSERT_EQ(8, file_size);
// Check that opening non-existent file fails. // Check that opening non-existent file fails.
@ -83,50 +85,49 @@ TEST_F(MemEnvTest, Basics) {
ASSERT_TRUE(!rand_file); ASSERT_TRUE(!rand_file);
// Check that deleting works. // Check that deleting works.
ASSERT_TRUE(!env_->RemoveFile("/dir/non_existent").ok()); ASSERT_TRUE(!env_->DeleteFile("/dir/non_existent").ok());
ASSERT_LEVELDB_OK(env_->RemoveFile("/dir/g")); ASSERT_OK(env_->DeleteFile("/dir/g"));
ASSERT_TRUE(!env_->FileExists("/dir/g")); ASSERT_TRUE(!env_->FileExists("/dir/g"));
ASSERT_LEVELDB_OK(env_->GetChildren("/dir", &children)); ASSERT_OK(env_->GetChildren("/dir", &children));
ASSERT_EQ(0, children.size()); ASSERT_EQ(0, children.size());
ASSERT_LEVELDB_OK(env_->RemoveDir("/dir")); ASSERT_OK(env_->DeleteDir("/dir"));
} }
TEST_F(MemEnvTest, ReadWrite) { TEST(MemEnvTest, ReadWrite) {
WritableFile* writable_file; WritableFile* writable_file;
SequentialFile* seq_file; SequentialFile* seq_file;
RandomAccessFile* rand_file; RandomAccessFile* rand_file;
Slice result; Slice result;
char scratch[100]; char scratch[100];
ASSERT_LEVELDB_OK(env_->CreateDir("/dir")); ASSERT_OK(env_->CreateDir("/dir"));
ASSERT_LEVELDB_OK(env_->NewWritableFile("/dir/f", &writable_file)); ASSERT_OK(env_->NewWritableFile("/dir/f", &writable_file));
ASSERT_LEVELDB_OK(writable_file->Append("hello ")); ASSERT_OK(writable_file->Append("hello "));
ASSERT_LEVELDB_OK(writable_file->Append("world")); ASSERT_OK(writable_file->Append("world"));
delete writable_file; delete writable_file;
// Read sequentially. // Read sequentially.
ASSERT_LEVELDB_OK(env_->NewSequentialFile("/dir/f", &seq_file)); ASSERT_OK(env_->NewSequentialFile("/dir/f", &seq_file));
ASSERT_LEVELDB_OK(seq_file->Read(5, &result, scratch)); // Read "hello". ASSERT_OK(seq_file->Read(5, &result, scratch)); // Read "hello".
ASSERT_EQ(0, result.compare("hello")); ASSERT_EQ(0, result.compare("hello"));
ASSERT_LEVELDB_OK(seq_file->Skip(1)); ASSERT_OK(seq_file->Skip(1));
ASSERT_LEVELDB_OK(seq_file->Read(1000, &result, scratch)); // Read "world". ASSERT_OK(seq_file->Read(1000, &result, scratch)); // Read "world".
ASSERT_EQ(0, result.compare("world")); ASSERT_EQ(0, result.compare("world"));
ASSERT_LEVELDB_OK( ASSERT_OK(seq_file->Read(1000, &result, scratch)); // Try reading past EOF.
seq_file->Read(1000, &result, scratch)); // Try reading past EOF.
ASSERT_EQ(0, result.size()); ASSERT_EQ(0, result.size());
ASSERT_LEVELDB_OK(seq_file->Skip(100)); // Try to skip past end of file. ASSERT_OK(seq_file->Skip(100)); // Try to skip past end of file.
ASSERT_LEVELDB_OK(seq_file->Read(1000, &result, scratch)); ASSERT_OK(seq_file->Read(1000, &result, scratch));
ASSERT_EQ(0, result.size()); ASSERT_EQ(0, result.size());
delete seq_file; delete seq_file;
// Random reads. // Random reads.
ASSERT_LEVELDB_OK(env_->NewRandomAccessFile("/dir/f", &rand_file)); ASSERT_OK(env_->NewRandomAccessFile("/dir/f", &rand_file));
ASSERT_LEVELDB_OK(rand_file->Read(6, 5, &result, scratch)); // Read "world". ASSERT_OK(rand_file->Read(6, 5, &result, scratch)); // Read "world".
ASSERT_EQ(0, result.compare("world")); ASSERT_EQ(0, result.compare("world"));
ASSERT_LEVELDB_OK(rand_file->Read(0, 5, &result, scratch)); // Read "hello". ASSERT_OK(rand_file->Read(0, 5, &result, scratch)); // Read "hello".
ASSERT_EQ(0, result.compare("hello")); ASSERT_EQ(0, result.compare("hello"));
ASSERT_LEVELDB_OK(rand_file->Read(10, 100, &result, scratch)); // Read "d". ASSERT_OK(rand_file->Read(10, 100, &result, scratch)); // Read "d".
ASSERT_EQ(0, result.compare("d")); ASSERT_EQ(0, result.compare("d"));
// Too high offset. // Too high offset.
@ -134,30 +135,30 @@ TEST_F(MemEnvTest, ReadWrite) {
delete rand_file; delete rand_file;
} }
TEST_F(MemEnvTest, Locks) { TEST(MemEnvTest, Locks) {
FileLock* lock; FileLock* lock;
// These are no-ops, but we test they return success. // These are no-ops, but we test they return success.
ASSERT_LEVELDB_OK(env_->LockFile("some file", &lock)); ASSERT_OK(env_->LockFile("some file", &lock));
ASSERT_LEVELDB_OK(env_->UnlockFile(lock)); ASSERT_OK(env_->UnlockFile(lock));
} }
TEST_F(MemEnvTest, Misc) { TEST(MemEnvTest, Misc) {
std::string test_dir; std::string test_dir;
ASSERT_LEVELDB_OK(env_->GetTestDirectory(&test_dir)); ASSERT_OK(env_->GetTestDirectory(&test_dir));
ASSERT_TRUE(!test_dir.empty()); ASSERT_TRUE(!test_dir.empty());
WritableFile* writable_file; WritableFile* writable_file;
ASSERT_LEVELDB_OK(env_->NewWritableFile("/a/b", &writable_file)); ASSERT_OK(env_->NewWritableFile("/a/b", &writable_file));
// These are no-ops, but we test they return success. // These are no-ops, but we test they return success.
ASSERT_LEVELDB_OK(writable_file->Sync()); ASSERT_OK(writable_file->Sync());
ASSERT_LEVELDB_OK(writable_file->Flush()); ASSERT_OK(writable_file->Flush());
ASSERT_LEVELDB_OK(writable_file->Close()); ASSERT_OK(writable_file->Close());
delete writable_file; delete writable_file;
} }
TEST_F(MemEnvTest, LargeWrite) { TEST(MemEnvTest, LargeWrite) {
const size_t kWriteSize = 300 * 1024; const size_t kWriteSize = 300 * 1024;
char* scratch = new char[kWriteSize * 2]; char* scratch = new char[kWriteSize * 2];
@ -167,53 +168,30 @@ TEST_F(MemEnvTest, LargeWrite) {
} }
WritableFile* writable_file; WritableFile* writable_file;
ASSERT_LEVELDB_OK(env_->NewWritableFile("/dir/f", &writable_file)); ASSERT_OK(env_->NewWritableFile("/dir/f", &writable_file));
ASSERT_LEVELDB_OK(writable_file->Append("foo")); ASSERT_OK(writable_file->Append("foo"));
ASSERT_LEVELDB_OK(writable_file->Append(write_data)); ASSERT_OK(writable_file->Append(write_data));
delete writable_file; delete writable_file;
SequentialFile* seq_file; SequentialFile* seq_file;
Slice result; Slice result;
ASSERT_LEVELDB_OK(env_->NewSequentialFile("/dir/f", &seq_file)); ASSERT_OK(env_->NewSequentialFile("/dir/f", &seq_file));
ASSERT_LEVELDB_OK(seq_file->Read(3, &result, scratch)); // Read "foo". ASSERT_OK(seq_file->Read(3, &result, scratch)); // Read "foo".
ASSERT_EQ(0, result.compare("foo")); ASSERT_EQ(0, result.compare("foo"));
size_t read = 0; size_t read = 0;
std::string read_data; std::string read_data;
while (read < kWriteSize) { while (read < kWriteSize) {
ASSERT_LEVELDB_OK(seq_file->Read(kWriteSize - read, &result, scratch)); ASSERT_OK(seq_file->Read(kWriteSize - read, &result, scratch));
read_data.append(result.data(), result.size()); read_data.append(result.data(), result.size());
read += result.size(); read += result.size();
} }
ASSERT_TRUE(write_data == read_data); ASSERT_TRUE(write_data == read_data);
delete seq_file; delete seq_file;
delete[] scratch; delete [] scratch;
} }
TEST_F(MemEnvTest, OverwriteOpenFile) { TEST(MemEnvTest, DBTest) {
const char kWrite1Data[] = "Write #1 data";
const size_t kFileDataLen = sizeof(kWrite1Data) - 1;
const std::string kTestFileName = testing::TempDir() + "leveldb-TestFile.dat";
ASSERT_LEVELDB_OK(WriteStringToFile(env_, kWrite1Data, kTestFileName));
RandomAccessFile* rand_file;
ASSERT_LEVELDB_OK(env_->NewRandomAccessFile(kTestFileName, &rand_file));
const char kWrite2Data[] = "Write #2 data";
ASSERT_LEVELDB_OK(WriteStringToFile(env_, kWrite2Data, kTestFileName));
// Verify that overwriting an open file will result in the new file data
// being read from files opened before the write.
Slice result;
char scratch[kFileDataLen];
ASSERT_LEVELDB_OK(rand_file->Read(0, kFileDataLen, &result, scratch));
ASSERT_EQ(0, result.compare(kWrite2Data));
delete rand_file;
}
TEST_F(MemEnvTest, DBTest) {
Options options; Options options;
options.create_if_missing = true; options.create_if_missing = true;
options.env = env_; options.env = env_;
@ -222,14 +200,14 @@ TEST_F(MemEnvTest, DBTest) {
const Slice keys[] = {Slice("aaa"), Slice("bbb"), Slice("ccc")}; const Slice keys[] = {Slice("aaa"), Slice("bbb"), Slice("ccc")};
const Slice vals[] = {Slice("foo"), Slice("bar"), Slice("baz")}; const Slice vals[] = {Slice("foo"), Slice("bar"), Slice("baz")};
ASSERT_LEVELDB_OK(DB::Open(options, "/dir/db", &db)); ASSERT_OK(DB::Open(options, "/dir/db", &db));
for (size_t i = 0; i < 3; ++i) { for (size_t i = 0; i < 3; ++i) {
ASSERT_LEVELDB_OK(db->Put(WriteOptions(), keys[i], vals[i])); ASSERT_OK(db->Put(WriteOptions(), keys[i], vals[i]));
} }
for (size_t i = 0; i < 3; ++i) { for (size_t i = 0; i < 3; ++i) {
std::string res; std::string res;
ASSERT_LEVELDB_OK(db->Get(ReadOptions(), keys[i], &res)); ASSERT_OK(db->Get(ReadOptions(), keys[i], &res));
ASSERT_TRUE(res == vals[i]); ASSERT_TRUE(res == vals[i]);
} }
@ -245,11 +223,11 @@ TEST_F(MemEnvTest, DBTest) {
delete iterator; delete iterator;
DBImpl* dbi = reinterpret_cast<DBImpl*>(db); DBImpl* dbi = reinterpret_cast<DBImpl*>(db);
ASSERT_LEVELDB_OK(dbi->TEST_CompactMemTable()); ASSERT_OK(dbi->TEST_CompactMemTable());
for (size_t i = 0; i < 3; ++i) { for (size_t i = 0; i < 3; ++i) {
std::string res; std::string res;
ASSERT_LEVELDB_OK(db->Get(ReadOptions(), keys[i], &res)); ASSERT_OK(db->Get(ReadOptions(), keys[i], &res));
ASSERT_TRUE(res == vals[i]); ASSERT_TRUE(res == vals[i]);
} }
@ -257,3 +235,7 @@ TEST_F(MemEnvTest, DBTest) {
} }
} // namespace leveldb } // namespace leveldb
int main(int argc, char** argv) {
return leveldb::test::RunAllTests();
}

View File

@ -32,7 +32,7 @@
On failure, leveldb frees the old value of *errptr and On failure, leveldb frees the old value of *errptr and
set *errptr to a malloc()ed error message. set *errptr to a malloc()ed error message.
(4) Bools have the type uint8_t (0 == false; rest == true) (4) Bools have the type unsigned char (0 == false; rest == true)
(5) All of the pointer arguments must be non-NULL. (5) All of the pointer arguments must be non-NULL.
*/ */
@ -40,213 +40,233 @@
#ifndef STORAGE_LEVELDB_INCLUDE_C_H_ #ifndef STORAGE_LEVELDB_INCLUDE_C_H_
#define STORAGE_LEVELDB_INCLUDE_C_H_ #define STORAGE_LEVELDB_INCLUDE_C_H_
#include <stdarg.h>
#include <stddef.h>
#include <stdint.h>
#include "leveldb/export.h"
#ifdef __cplusplus #ifdef __cplusplus
extern "C" { extern "C" {
#endif #endif
#include <stdarg.h>
#include <stddef.h>
#include <stdint.h>
/* Exported types */ /* Exported types */
typedef struct leveldb_t leveldb_t; typedef struct leveldb_t leveldb_t;
typedef struct leveldb_cache_t leveldb_cache_t; typedef struct leveldb_cache_t leveldb_cache_t;
typedef struct leveldb_comparator_t leveldb_comparator_t; typedef struct leveldb_comparator_t leveldb_comparator_t;
typedef struct leveldb_env_t leveldb_env_t; typedef struct leveldb_env_t leveldb_env_t;
typedef struct leveldb_filelock_t leveldb_filelock_t; typedef struct leveldb_filelock_t leveldb_filelock_t;
typedef struct leveldb_filterpolicy_t leveldb_filterpolicy_t; typedef struct leveldb_filterpolicy_t leveldb_filterpolicy_t;
typedef struct leveldb_iterator_t leveldb_iterator_t; typedef struct leveldb_iterator_t leveldb_iterator_t;
typedef struct leveldb_logger_t leveldb_logger_t; typedef struct leveldb_logger_t leveldb_logger_t;
typedef struct leveldb_options_t leveldb_options_t; typedef struct leveldb_options_t leveldb_options_t;
typedef struct leveldb_randomfile_t leveldb_randomfile_t; typedef struct leveldb_randomfile_t leveldb_randomfile_t;
typedef struct leveldb_readoptions_t leveldb_readoptions_t; typedef struct leveldb_readoptions_t leveldb_readoptions_t;
typedef struct leveldb_seqfile_t leveldb_seqfile_t; typedef struct leveldb_seqfile_t leveldb_seqfile_t;
typedef struct leveldb_snapshot_t leveldb_snapshot_t; typedef struct leveldb_snapshot_t leveldb_snapshot_t;
typedef struct leveldb_writablefile_t leveldb_writablefile_t; typedef struct leveldb_writablefile_t leveldb_writablefile_t;
typedef struct leveldb_writebatch_t leveldb_writebatch_t; typedef struct leveldb_writebatch_t leveldb_writebatch_t;
typedef struct leveldb_writeoptions_t leveldb_writeoptions_t; typedef struct leveldb_writeoptions_t leveldb_writeoptions_t;
/* DB operations */ /* DB operations */
LEVELDB_EXPORT leveldb_t* leveldb_open(const leveldb_options_t* options, extern leveldb_t* leveldb_open(
const char* name, char** errptr); const leveldb_options_t* options,
const char* name,
char** errptr);
LEVELDB_EXPORT void leveldb_close(leveldb_t* db); extern void leveldb_close(leveldb_t* db);
LEVELDB_EXPORT void leveldb_put(leveldb_t* db, extern void leveldb_put(
const leveldb_writeoptions_t* options, leveldb_t* db,
const char* key, size_t keylen, const char* val, const leveldb_writeoptions_t* options,
size_t vallen, char** errptr); const char* key, size_t keylen,
const char* val, size_t vallen,
char** errptr);
LEVELDB_EXPORT void leveldb_delete(leveldb_t* db, extern void leveldb_delete(
const leveldb_writeoptions_t* options, leveldb_t* db,
const char* key, size_t keylen, const leveldb_writeoptions_t* options,
char** errptr); const char* key, size_t keylen,
char** errptr);
LEVELDB_EXPORT void leveldb_write(leveldb_t* db, extern void leveldb_write(
const leveldb_writeoptions_t* options, leveldb_t* db,
leveldb_writebatch_t* batch, char** errptr); const leveldb_writeoptions_t* options,
leveldb_writebatch_t* batch,
char** errptr);
/* Returns NULL if not found. A malloc()ed array otherwise. /* Returns NULL if not found. A malloc()ed array otherwise.
Stores the length of the array in *vallen. */ Stores the length of the array in *vallen. */
LEVELDB_EXPORT char* leveldb_get(leveldb_t* db, extern char* leveldb_get(
const leveldb_readoptions_t* options, leveldb_t* db,
const char* key, size_t keylen, size_t* vallen, const leveldb_readoptions_t* options,
char** errptr); const char* key, size_t keylen,
size_t* vallen,
char** errptr);
LEVELDB_EXPORT leveldb_iterator_t* leveldb_create_iterator( extern leveldb_iterator_t* leveldb_create_iterator(
leveldb_t* db, const leveldb_readoptions_t* options); leveldb_t* db,
const leveldb_readoptions_t* options);
LEVELDB_EXPORT const leveldb_snapshot_t* leveldb_create_snapshot(leveldb_t* db); extern const leveldb_snapshot_t* leveldb_create_snapshot(
leveldb_t* db);
LEVELDB_EXPORT void leveldb_release_snapshot( extern void leveldb_release_snapshot(
leveldb_t* db, const leveldb_snapshot_t* snapshot); leveldb_t* db,
const leveldb_snapshot_t* snapshot);
/* Returns NULL if property name is unknown. /* Returns NULL if property name is unknown.
Else returns a pointer to a malloc()-ed null-terminated value. */ Else returns a pointer to a malloc()-ed null-terminated value. */
LEVELDB_EXPORT char* leveldb_property_value(leveldb_t* db, extern char* leveldb_property_value(
const char* propname); leveldb_t* db,
const char* propname);
LEVELDB_EXPORT void leveldb_approximate_sizes( extern void leveldb_approximate_sizes(
leveldb_t* db, int num_ranges, const char* const* range_start_key, leveldb_t* db,
const size_t* range_start_key_len, const char* const* range_limit_key, int num_ranges,
const size_t* range_limit_key_len, uint64_t* sizes); const char* const* range_start_key, const size_t* range_start_key_len,
const char* const* range_limit_key, const size_t* range_limit_key_len,
uint64_t* sizes);
LEVELDB_EXPORT void leveldb_compact_range(leveldb_t* db, const char* start_key, extern void leveldb_compact_range(
size_t start_key_len, leveldb_t* db,
const char* limit_key, const char* start_key, size_t start_key_len,
size_t limit_key_len); const char* limit_key, size_t limit_key_len);
/* Management operations */ /* Management operations */
LEVELDB_EXPORT void leveldb_destroy_db(const leveldb_options_t* options, extern void leveldb_destroy_db(
const char* name, char** errptr); const leveldb_options_t* options,
const char* name,
char** errptr);
LEVELDB_EXPORT void leveldb_repair_db(const leveldb_options_t* options, extern void leveldb_repair_db(
const char* name, char** errptr); const leveldb_options_t* options,
const char* name,
char** errptr);
/* Iterator */ /* Iterator */
LEVELDB_EXPORT void leveldb_iter_destroy(leveldb_iterator_t*); extern void leveldb_iter_destroy(leveldb_iterator_t*);
LEVELDB_EXPORT uint8_t leveldb_iter_valid(const leveldb_iterator_t*); extern unsigned char leveldb_iter_valid(const leveldb_iterator_t*);
LEVELDB_EXPORT void leveldb_iter_seek_to_first(leveldb_iterator_t*); extern void leveldb_iter_seek_to_first(leveldb_iterator_t*);
LEVELDB_EXPORT void leveldb_iter_seek_to_last(leveldb_iterator_t*); extern void leveldb_iter_seek_to_last(leveldb_iterator_t*);
LEVELDB_EXPORT void leveldb_iter_seek(leveldb_iterator_t*, const char* k, extern void leveldb_iter_seek(leveldb_iterator_t*, const char* k, size_t klen);
size_t klen); extern void leveldb_iter_next(leveldb_iterator_t*);
LEVELDB_EXPORT void leveldb_iter_next(leveldb_iterator_t*); extern void leveldb_iter_prev(leveldb_iterator_t*);
LEVELDB_EXPORT void leveldb_iter_prev(leveldb_iterator_t*); extern const char* leveldb_iter_key(const leveldb_iterator_t*, size_t* klen);
LEVELDB_EXPORT const char* leveldb_iter_key(const leveldb_iterator_t*, extern const char* leveldb_iter_value(const leveldb_iterator_t*, size_t* vlen);
size_t* klen); extern void leveldb_iter_get_error(const leveldb_iterator_t*, char** errptr);
LEVELDB_EXPORT const char* leveldb_iter_value(const leveldb_iterator_t*,
size_t* vlen);
LEVELDB_EXPORT void leveldb_iter_get_error(const leveldb_iterator_t*,
char** errptr);
/* Write batch */ /* Write batch */
LEVELDB_EXPORT leveldb_writebatch_t* leveldb_writebatch_create(void); extern leveldb_writebatch_t* leveldb_writebatch_create();
LEVELDB_EXPORT void leveldb_writebatch_destroy(leveldb_writebatch_t*); extern void leveldb_writebatch_destroy(leveldb_writebatch_t*);
LEVELDB_EXPORT void leveldb_writebatch_clear(leveldb_writebatch_t*); extern void leveldb_writebatch_clear(leveldb_writebatch_t*);
LEVELDB_EXPORT void leveldb_writebatch_put(leveldb_writebatch_t*, extern void leveldb_writebatch_put(
const char* key, size_t klen, leveldb_writebatch_t*,
const char* val, size_t vlen); const char* key, size_t klen,
LEVELDB_EXPORT void leveldb_writebatch_delete(leveldb_writebatch_t*, const char* val, size_t vlen);
const char* key, size_t klen); extern void leveldb_writebatch_delete(
LEVELDB_EXPORT void leveldb_writebatch_iterate( leveldb_writebatch_t*,
const leveldb_writebatch_t*, void* state, const char* key, size_t klen);
extern void leveldb_writebatch_iterate(
leveldb_writebatch_t*,
void* state,
void (*put)(void*, const char* k, size_t klen, const char* v, size_t vlen), void (*put)(void*, const char* k, size_t klen, const char* v, size_t vlen),
void (*deleted)(void*, const char* k, size_t klen)); void (*deleted)(void*, const char* k, size_t klen));
LEVELDB_EXPORT void leveldb_writebatch_append(
leveldb_writebatch_t* destination, const leveldb_writebatch_t* source);
/* Options */ /* Options */
LEVELDB_EXPORT leveldb_options_t* leveldb_options_create(void); extern leveldb_options_t* leveldb_options_create();
LEVELDB_EXPORT void leveldb_options_destroy(leveldb_options_t*); extern void leveldb_options_destroy(leveldb_options_t*);
LEVELDB_EXPORT void leveldb_options_set_comparator(leveldb_options_t*, extern void leveldb_options_set_comparator(
leveldb_comparator_t*); leveldb_options_t*,
LEVELDB_EXPORT void leveldb_options_set_filter_policy(leveldb_options_t*, leveldb_comparator_t*);
leveldb_filterpolicy_t*); extern void leveldb_options_set_filter_policy(
LEVELDB_EXPORT void leveldb_options_set_create_if_missing(leveldb_options_t*, leveldb_options_t*,
uint8_t); leveldb_filterpolicy_t*);
LEVELDB_EXPORT void leveldb_options_set_error_if_exists(leveldb_options_t*, extern void leveldb_options_set_create_if_missing(
uint8_t); leveldb_options_t*, unsigned char);
LEVELDB_EXPORT void leveldb_options_set_paranoid_checks(leveldb_options_t*, extern void leveldb_options_set_error_if_exists(
uint8_t); leveldb_options_t*, unsigned char);
LEVELDB_EXPORT void leveldb_options_set_env(leveldb_options_t*, leveldb_env_t*); extern void leveldb_options_set_paranoid_checks(
LEVELDB_EXPORT void leveldb_options_set_info_log(leveldb_options_t*, leveldb_options_t*, unsigned char);
leveldb_logger_t*); extern void leveldb_options_set_env(leveldb_options_t*, leveldb_env_t*);
LEVELDB_EXPORT void leveldb_options_set_write_buffer_size(leveldb_options_t*, extern void leveldb_options_set_info_log(leveldb_options_t*, leveldb_logger_t*);
size_t); extern void leveldb_options_set_write_buffer_size(leveldb_options_t*, size_t);
LEVELDB_EXPORT void leveldb_options_set_max_open_files(leveldb_options_t*, int); extern void leveldb_options_set_max_open_files(leveldb_options_t*, int);
LEVELDB_EXPORT void leveldb_options_set_cache(leveldb_options_t*, extern void leveldb_options_set_cache(leveldb_options_t*, leveldb_cache_t*);
leveldb_cache_t*); extern void leveldb_options_set_block_size(leveldb_options_t*, size_t);
LEVELDB_EXPORT void leveldb_options_set_block_size(leveldb_options_t*, size_t); extern void leveldb_options_set_block_restart_interval(leveldb_options_t*, int);
LEVELDB_EXPORT void leveldb_options_set_block_restart_interval(
leveldb_options_t*, int);
LEVELDB_EXPORT void leveldb_options_set_max_file_size(leveldb_options_t*,
size_t);
enum { leveldb_no_compression = 0, leveldb_snappy_compression = 1 }; enum {
LEVELDB_EXPORT void leveldb_options_set_compression(leveldb_options_t*, int); leveldb_no_compression = 0,
leveldb_snappy_compression = 1
};
extern void leveldb_options_set_compression(leveldb_options_t*, int);
/* Comparator */ /* Comparator */
LEVELDB_EXPORT leveldb_comparator_t* leveldb_comparator_create( extern leveldb_comparator_t* leveldb_comparator_create(
void* state, void (*destructor)(void*), void* state,
int (*compare)(void*, const char* a, size_t alen, const char* b, void (*destructor)(void*),
size_t blen), int (*compare)(
void*,
const char* a, size_t alen,
const char* b, size_t blen),
const char* (*name)(void*)); const char* (*name)(void*));
LEVELDB_EXPORT void leveldb_comparator_destroy(leveldb_comparator_t*); extern void leveldb_comparator_destroy(leveldb_comparator_t*);
/* Filter policy */ /* Filter policy */
LEVELDB_EXPORT leveldb_filterpolicy_t* leveldb_filterpolicy_create( extern leveldb_filterpolicy_t* leveldb_filterpolicy_create(
void* state, void (*destructor)(void*), void* state,
char* (*create_filter)(void*, const char* const* key_array, void (*destructor)(void*),
const size_t* key_length_array, int num_keys, char* (*create_filter)(
size_t* filter_length), void*,
uint8_t (*key_may_match)(void*, const char* key, size_t length, const char* const* key_array, const size_t* key_length_array,
const char* filter, size_t filter_length), int num_keys,
size_t* filter_length),
unsigned char (*key_may_match)(
void*,
const char* key, size_t length,
const char* filter, size_t filter_length),
const char* (*name)(void*)); const char* (*name)(void*));
LEVELDB_EXPORT void leveldb_filterpolicy_destroy(leveldb_filterpolicy_t*); extern void leveldb_filterpolicy_destroy(leveldb_filterpolicy_t*);
LEVELDB_EXPORT leveldb_filterpolicy_t* leveldb_filterpolicy_create_bloom( extern leveldb_filterpolicy_t* leveldb_filterpolicy_create_bloom(
int bits_per_key); int bits_per_key);
/* Read options */ /* Read options */
LEVELDB_EXPORT leveldb_readoptions_t* leveldb_readoptions_create(void); extern leveldb_readoptions_t* leveldb_readoptions_create();
LEVELDB_EXPORT void leveldb_readoptions_destroy(leveldb_readoptions_t*); extern void leveldb_readoptions_destroy(leveldb_readoptions_t*);
LEVELDB_EXPORT void leveldb_readoptions_set_verify_checksums( extern void leveldb_readoptions_set_verify_checksums(
leveldb_readoptions_t*, uint8_t); leveldb_readoptions_t*,
LEVELDB_EXPORT void leveldb_readoptions_set_fill_cache(leveldb_readoptions_t*, unsigned char);
uint8_t); extern void leveldb_readoptions_set_fill_cache(
LEVELDB_EXPORT void leveldb_readoptions_set_snapshot(leveldb_readoptions_t*, leveldb_readoptions_t*, unsigned char);
const leveldb_snapshot_t*); extern void leveldb_readoptions_set_snapshot(
leveldb_readoptions_t*,
const leveldb_snapshot_t*);
/* Write options */ /* Write options */
LEVELDB_EXPORT leveldb_writeoptions_t* leveldb_writeoptions_create(void); extern leveldb_writeoptions_t* leveldb_writeoptions_create();
LEVELDB_EXPORT void leveldb_writeoptions_destroy(leveldb_writeoptions_t*); extern void leveldb_writeoptions_destroy(leveldb_writeoptions_t*);
LEVELDB_EXPORT void leveldb_writeoptions_set_sync(leveldb_writeoptions_t*, extern void leveldb_writeoptions_set_sync(
uint8_t); leveldb_writeoptions_t*, unsigned char);
/* Cache */ /* Cache */
LEVELDB_EXPORT leveldb_cache_t* leveldb_cache_create_lru(size_t capacity); extern leveldb_cache_t* leveldb_cache_create_lru(size_t capacity);
LEVELDB_EXPORT void leveldb_cache_destroy(leveldb_cache_t* cache); extern void leveldb_cache_destroy(leveldb_cache_t* cache);
/* Env */ /* Env */
LEVELDB_EXPORT leveldb_env_t* leveldb_create_default_env(void); extern leveldb_env_t* leveldb_create_default_env();
LEVELDB_EXPORT void leveldb_env_destroy(leveldb_env_t*); extern void leveldb_env_destroy(leveldb_env_t*);
/* If not NULL, the returned buffer must be released using leveldb_free(). */
LEVELDB_EXPORT char* leveldb_env_get_test_directory(leveldb_env_t*);
/* Utility */ /* Utility */
@ -255,16 +275,16 @@ LEVELDB_EXPORT char* leveldb_env_get_test_directory(leveldb_env_t*);
in this file. Note that in certain cases (typically on Windows), you in this file. Note that in certain cases (typically on Windows), you
may need to call this routine instead of free(ptr) to dispose of may need to call this routine instead of free(ptr) to dispose of
malloc()-ed memory returned by this library. */ malloc()-ed memory returned by this library. */
LEVELDB_EXPORT void leveldb_free(void* ptr); extern void leveldb_free(void* ptr);
/* Return the major version number for this release. */ /* Return the major version number for this release. */
LEVELDB_EXPORT int leveldb_major_version(void); extern int leveldb_major_version();
/* Return the minor version number for this release. */ /* Return the minor version number for this release. */
LEVELDB_EXPORT int leveldb_minor_version(void); extern int leveldb_minor_version();
#ifdef __cplusplus #ifdef __cplusplus
} /* end extern "C" */ } /* end extern "C" */
#endif #endif
#endif /* STORAGE_LEVELDB_INCLUDE_C_H_ */ #endif /* STORAGE_LEVELDB_INCLUDE_C_H_ */

View File

@ -18,32 +18,27 @@
#ifndef STORAGE_LEVELDB_INCLUDE_CACHE_H_ #ifndef STORAGE_LEVELDB_INCLUDE_CACHE_H_
#define STORAGE_LEVELDB_INCLUDE_CACHE_H_ #define STORAGE_LEVELDB_INCLUDE_CACHE_H_
#include <cstdint> #include <stdint.h>
#include "leveldb/export.h"
#include "leveldb/slice.h" #include "leveldb/slice.h"
namespace leveldb { namespace leveldb {
class LEVELDB_EXPORT Cache; class Cache;
// Create a new cache with a fixed size capacity. This implementation // Create a new cache with a fixed size capacity. This implementation
// of Cache uses a least-recently-used eviction policy. // of Cache uses a least-recently-used eviction policy.
LEVELDB_EXPORT Cache* NewLRUCache(size_t capacity); extern Cache* NewLRUCache(size_t capacity);
class LEVELDB_EXPORT Cache { class Cache {
public: public:
Cache() = default; Cache() { }
Cache(const Cache&) = delete;
Cache& operator=(const Cache&) = delete;
// Destroys all existing entries by calling the "deleter" // Destroys all existing entries by calling the "deleter"
// function that was passed to the constructor. // function that was passed to the constructor.
virtual ~Cache(); virtual ~Cache();
// Opaque handle to an entry stored in the cache. // Opaque handle to an entry stored in the cache.
struct Handle {}; struct Handle { };
// Insert a mapping from key->value into the cache and assign it // Insert a mapping from key->value into the cache and assign it
// the specified charge against the total cache capacity. // the specified charge against the total cache capacity.
@ -57,7 +52,7 @@ class LEVELDB_EXPORT Cache {
virtual Handle* Insert(const Slice& key, void* value, size_t charge, virtual Handle* Insert(const Slice& key, void* value, size_t charge,
void (*deleter)(const Slice& key, void* value)) = 0; void (*deleter)(const Slice& key, void* value)) = 0;
// If the cache has no mapping for "key", returns nullptr. // If the cache has no mapping for "key", returns NULL.
// //
// Else return a handle that corresponds to the mapping. The caller // Else return a handle that corresponds to the mapping. The caller
// must call this->Release(handle) when the returned mapping is no // must call this->Release(handle) when the returned mapping is no
@ -96,6 +91,18 @@ class LEVELDB_EXPORT Cache {
// Return an estimate of the combined charges of all elements stored in the // Return an estimate of the combined charges of all elements stored in the
// cache. // cache.
virtual size_t TotalCharge() const = 0; virtual size_t TotalCharge() const = 0;
private:
void LRU_Remove(Handle* e);
void LRU_Append(Handle* e);
void Unref(Handle* e);
struct Rep;
Rep* rep_;
// No copying allowed
Cache(const Cache&);
void operator=(const Cache&);
}; };
} // namespace leveldb } // namespace leveldb

View File

@ -7,8 +7,6 @@
#include <string> #include <string>
#include "leveldb/export.h"
namespace leveldb { namespace leveldb {
class Slice; class Slice;
@ -17,7 +15,7 @@ class Slice;
// used as keys in an sstable or a database. A Comparator implementation // used as keys in an sstable or a database. A Comparator implementation
// must be thread-safe since leveldb may invoke its methods concurrently // must be thread-safe since leveldb may invoke its methods concurrently
// from multiple threads. // from multiple threads.
class LEVELDB_EXPORT Comparator { class Comparator {
public: public:
virtual ~Comparator(); virtual ~Comparator();
@ -45,8 +43,9 @@ class LEVELDB_EXPORT Comparator {
// If *start < limit, changes *start to a short string in [start,limit). // If *start < limit, changes *start to a short string in [start,limit).
// Simple comparator implementations may return with *start unchanged, // Simple comparator implementations may return with *start unchanged,
// i.e., an implementation of this method that does nothing is correct. // i.e., an implementation of this method that does nothing is correct.
virtual void FindShortestSeparator(std::string* start, virtual void FindShortestSeparator(
const Slice& limit) const = 0; std::string* start,
const Slice& limit) const = 0;
// Changes *key to a short string >= *key. // Changes *key to a short string >= *key.
// Simple comparator implementations may return with *key unchanged, // Simple comparator implementations may return with *key unchanged,
@ -57,7 +56,7 @@ class LEVELDB_EXPORT Comparator {
// Return a builtin comparator that uses lexicographic byte-wise // Return a builtin comparator that uses lexicographic byte-wise
// ordering. The result remains the property of this module and // ordering. The result remains the property of this module and
// must not be deleted. // must not be deleted.
LEVELDB_EXPORT const Comparator* BytewiseComparator(); extern const Comparator* BytewiseComparator();
} // namespace leveldb } // namespace leveldb

View File

@ -5,18 +5,16 @@
#ifndef STORAGE_LEVELDB_INCLUDE_DB_H_ #ifndef STORAGE_LEVELDB_INCLUDE_DB_H_
#define STORAGE_LEVELDB_INCLUDE_DB_H_ #define STORAGE_LEVELDB_INCLUDE_DB_H_
#include <cstdint> #include <stdint.h>
#include <cstdio> #include <stdio.h>
#include "leveldb/export.h"
#include "leveldb/iterator.h" #include "leveldb/iterator.h"
#include "leveldb/options.h" #include "leveldb/options.h"
namespace leveldb { namespace leveldb {
// Update CMakeLists.txt if you change these // Update Makefile if you change these
static const int kMajorVersion = 1; static const int kMajorVersion = 1;
static const int kMinorVersion = 23; static const int kMinorVersion = 18;
struct Options; struct Options;
struct ReadOptions; struct ReadOptions;
@ -26,44 +24,42 @@ class WriteBatch;
// Abstract handle to particular state of a DB. // Abstract handle to particular state of a DB.
// A Snapshot is an immutable object and can therefore be safely // A Snapshot is an immutable object and can therefore be safely
// accessed from multiple threads without any external synchronization. // accessed from multiple threads without any external synchronization.
class LEVELDB_EXPORT Snapshot { class Snapshot {
protected: protected:
virtual ~Snapshot(); virtual ~Snapshot();
}; };
// A range of keys // A range of keys
struct LEVELDB_EXPORT Range { struct Range {
Range() = default; Slice start; // Included in the range
Range(const Slice& s, const Slice& l) : start(s), limit(l) {} Slice limit; // Not included in the range
Slice start; // Included in the range Range() { }
Slice limit; // Not included in the range Range(const Slice& s, const Slice& l) : start(s), limit(l) { }
}; };
// A DB is a persistent ordered map from keys to values. // A DB is a persistent ordered map from keys to values.
// A DB is safe for concurrent access from multiple threads without // A DB is safe for concurrent access from multiple threads without
// any external synchronization. // any external synchronization.
class LEVELDB_EXPORT DB { class DB {
public: public:
// Open the database with the specified "name". // Open the database with the specified "name".
// Stores a pointer to a heap-allocated database in *dbptr and returns // Stores a pointer to a heap-allocated database in *dbptr and returns
// OK on success. // OK on success.
// Stores nullptr in *dbptr and returns a non-OK status on error. // Stores NULL in *dbptr and returns a non-OK status on error.
// Caller should delete *dbptr when it is no longer needed. // Caller should delete *dbptr when it is no longer needed.
static Status Open(const Options& options, const std::string& name, static Status Open(const Options& options,
const std::string& name,
DB** dbptr); DB** dbptr);
DB() = default; DB() { }
DB(const DB&) = delete;
DB& operator=(const DB&) = delete;
virtual ~DB(); virtual ~DB();
// Set the database entry for "key" to "value". Returns OK on success, // Set the database entry for "key" to "value". Returns OK on success,
// and a non-OK status on error. // and a non-OK status on error.
// Note: consider setting options.sync = true. // Note: consider setting options.sync = true.
virtual Status Put(const WriteOptions& options, const Slice& key, virtual Status Put(const WriteOptions& options,
const Slice& key,
const Slice& value) = 0; const Slice& value) = 0;
// Remove the database entry (if any) for "key". Returns OK on // Remove the database entry (if any) for "key". Returns OK on
@ -84,8 +80,8 @@ class LEVELDB_EXPORT DB {
// a status for which Status::IsNotFound() returns true. // a status for which Status::IsNotFound() returns true.
// //
// May return some other Status on an error. // May return some other Status on an error.
virtual Status Get(const ReadOptions& options, const Slice& key, virtual Status Get(const ReadOptions& options,
std::string* value) = 0; const Slice& key, std::string* value) = 0;
// Return a heap-allocated iterator over the contents of the database. // Return a heap-allocated iterator over the contents of the database.
// The result of NewIterator() is initially invalid (caller must // The result of NewIterator() is initially invalid (caller must
@ -140,27 +136,27 @@ class LEVELDB_EXPORT DB {
// needed to access the data. This operation should typically only // needed to access the data. This operation should typically only
// be invoked by users who understand the underlying implementation. // be invoked by users who understand the underlying implementation.
// //
// begin==nullptr is treated as a key before all keys in the database. // begin==NULL is treated as a key before all keys in the database.
// end==nullptr is treated as a key after all keys in the database. // end==NULL is treated as a key after all keys in the database.
// Therefore the following call will compact the entire database: // Therefore the following call will compact the entire database:
// db->CompactRange(nullptr, nullptr); // db->CompactRange(NULL, NULL);
virtual void CompactRange(const Slice* begin, const Slice* end) = 0; virtual void CompactRange(const Slice* begin, const Slice* end) = 0;
private:
// No copying allowed
DB(const DB&);
void operator=(const DB&);
}; };
// Destroy the contents of the specified database. // Destroy the contents of the specified database.
// Be very careful using this method. // Be very careful using this method.
// Status DestroyDB(const std::string& name, const Options& options);
// Note: For backwards compatibility, if DestroyDB is unable to list the
// database files, Status::OK() will still be returned masking this failure.
LEVELDB_EXPORT Status DestroyDB(const std::string& name,
const Options& options);
// If a DB cannot be opened, you may attempt to call this method to // If a DB cannot be opened, you may attempt to call this method to
// resurrect as much of the contents of the database as possible. // resurrect as much of the contents of the database as possible.
// Some data may be lost, so be careful when calling this function // Some data may be lost, so be careful when calling this function
// on a database that contains important information. // on a database that contains important information.
LEVELDB_EXPORT Status RepairDB(const std::string& dbname, Status RepairDB(const std::string& dbname, const Options& options);
const Options& options);
} // namespace leveldb } // namespace leveldb

View File

@ -6,9 +6,7 @@
#define STORAGE_LEVELDB_INCLUDE_DUMPFILE_H_ #define STORAGE_LEVELDB_INCLUDE_DUMPFILE_H_
#include <string> #include <string>
#include "leveldb/env.h" #include "leveldb/env.h"
#include "leveldb/export.h"
#include "leveldb/status.h" #include "leveldb/status.h"
namespace leveldb { namespace leveldb {
@ -20,8 +18,7 @@ namespace leveldb {
// //
// Returns a non-OK result if fname does not name a leveldb storage // Returns a non-OK result if fname does not name a leveldb storage
// file, or if the file cannot be read. // file, or if the file cannot be read.
LEVELDB_EXPORT Status DumpFile(Env* env, const std::string& fname, Status DumpFile(Env* env, const std::string& fname, WritableFile* dst);
WritableFile* dst);
} // namespace leveldb } // namespace leveldb

View File

@ -13,32 +13,12 @@
#ifndef STORAGE_LEVELDB_INCLUDE_ENV_H_ #ifndef STORAGE_LEVELDB_INCLUDE_ENV_H_
#define STORAGE_LEVELDB_INCLUDE_ENV_H_ #define STORAGE_LEVELDB_INCLUDE_ENV_H_
#include <cstdarg>
#include <cstdint>
#include <string> #include <string>
#include <vector> #include <vector>
#include <stdarg.h>
#include "leveldb/export.h" #include <stdint.h>
#include "leveldb/status.h" #include "leveldb/status.h"
// This workaround can be removed when leveldb::Env::DeleteFile is removed.
#if defined(_WIN32)
// On Windows, the method name DeleteFile (below) introduces the risk of
// triggering undefined behavior by exposing the compiler to different
// declarations of the Env class in different translation units.
//
// This is because <windows.h>, a fairly popular header file for Windows
// applications, defines a DeleteFile macro. So, files that include the Windows
// header before this header will contain an altered Env declaration.
//
// This workaround ensures that the compiler sees the same Env declaration,
// independently of whether <windows.h> was included.
#if defined(DeleteFile)
#undef DeleteFile
#define LEVELDB_DELETEFILE_UNDEFINED
#endif // defined(DeleteFile)
#endif // defined(_WIN32)
namespace leveldb { namespace leveldb {
class FileLock; class FileLock;
@ -48,13 +28,9 @@ class SequentialFile;
class Slice; class Slice;
class WritableFile; class WritableFile;
class LEVELDB_EXPORT Env { class Env {
public: public:
Env(); Env() { }
Env(const Env&) = delete;
Env& operator=(const Env&) = delete;
virtual ~Env(); virtual ~Env();
// Return a default environment suitable for the current operating // Return a default environment suitable for the current operating
@ -64,22 +40,20 @@ class LEVELDB_EXPORT Env {
// The result of Default() belongs to leveldb and must never be deleted. // The result of Default() belongs to leveldb and must never be deleted.
static Env* Default(); static Env* Default();
// Create an object that sequentially reads the file with the specified name. // Create a brand new sequentially-readable file with the specified name.
// On success, stores a pointer to the new file in *result and returns OK. // On success, stores a pointer to the new file in *result and returns OK.
// On failure stores nullptr in *result and returns non-OK. If the file does // On failure stores NULL in *result and returns non-OK. If the file does
// not exist, returns a non-OK status. Implementations should return a // not exist, returns a non-OK status.
// NotFound status when the file does not exist.
// //
// The returned file will only be accessed by one thread at a time. // The returned file will only be accessed by one thread at a time.
virtual Status NewSequentialFile(const std::string& fname, virtual Status NewSequentialFile(const std::string& fname,
SequentialFile** result) = 0; SequentialFile** result) = 0;
// Create an object supporting random-access reads from the file with the // Create a brand new random access read-only file with the
// specified name. On success, stores a pointer to the new file in // specified name. On success, stores a pointer to the new file in
// *result and returns OK. On failure stores nullptr in *result and // *result and returns OK. On failure stores NULL in *result and
// returns non-OK. If the file does not exist, returns a non-OK // returns non-OK. If the file does not exist, returns a non-OK
// status. Implementations should return a NotFound status when the file does // status.
// not exist.
// //
// The returned file may be concurrently accessed by multiple threads. // The returned file may be concurrently accessed by multiple threads.
virtual Status NewRandomAccessFile(const std::string& fname, virtual Status NewRandomAccessFile(const std::string& fname,
@ -88,7 +62,7 @@ class LEVELDB_EXPORT Env {
// Create an object that writes to a new file with the specified // Create an object that writes to a new file with the specified
// name. Deletes any existing file with the same name and creates a // name. Deletes any existing file with the same name and creates a
// new file. On success, stores a pointer to the new file in // new file. On success, stores a pointer to the new file in
// *result and returns OK. On failure stores nullptr in *result and // *result and returns OK. On failure stores NULL in *result and
// returns non-OK. // returns non-OK.
// //
// The returned file will only be accessed by one thread at a time. // The returned file will only be accessed by one thread at a time.
@ -98,7 +72,7 @@ class LEVELDB_EXPORT Env {
// Create an object that either appends to an existing file, or // Create an object that either appends to an existing file, or
// writes to a new file (if the file does not exist to begin with). // writes to a new file (if the file does not exist to begin with).
// On success, stores a pointer to the new file in *result and // On success, stores a pointer to the new file in *result and
// returns OK. On failure stores nullptr in *result and returns // returns OK. On failure stores NULL in *result and returns
// non-OK. // non-OK.
// //
// The returned file will only be accessed by one thread at a time. // The returned file will only be accessed by one thread at a time.
@ -118,48 +92,15 @@ class LEVELDB_EXPORT Env {
// Original contents of *results are dropped. // Original contents of *results are dropped.
virtual Status GetChildren(const std::string& dir, virtual Status GetChildren(const std::string& dir,
std::vector<std::string>* result) = 0; std::vector<std::string>* result) = 0;
// Delete the named file.
//
// The default implementation calls DeleteFile, to support legacy Env
// implementations. Updated Env implementations must override RemoveFile and
// ignore the existence of DeleteFile. Updated code calling into the Env API
// must call RemoveFile instead of DeleteFile.
//
// A future release will remove DeleteDir and the default implementation of
// RemoveDir.
virtual Status RemoveFile(const std::string& fname);
// DEPRECATED: Modern Env implementations should override RemoveFile instead. // Delete the named file.
// virtual Status DeleteFile(const std::string& fname) = 0;
// The default implementation calls RemoveFile, to support legacy Env user
// code that calls this method on modern Env implementations. Modern Env user
// code should call RemoveFile.
//
// A future release will remove this method.
virtual Status DeleteFile(const std::string& fname);
// Create the specified directory. // Create the specified directory.
virtual Status CreateDir(const std::string& dirname) = 0; virtual Status CreateDir(const std::string& dirname) = 0;
// Delete the specified directory. // Delete the specified directory.
// virtual Status DeleteDir(const std::string& dirname) = 0;
// The default implementation calls DeleteDir, to support legacy Env
// implementations. Updated Env implementations must override RemoveDir and
// ignore the existence of DeleteDir. Modern code calling into the Env API
// must call RemoveDir instead of DeleteDir.
//
// A future release will remove DeleteDir and the default implementation of
// RemoveDir.
virtual Status RemoveDir(const std::string& dirname);
// DEPRECATED: Modern Env implementations should override RemoveDir instead.
//
// The default implementation calls RemoveDir, to support legacy Env user
// code that calls this method on modern Env implementations. Modern Env user
// code should call RemoveDir.
//
// A future release will remove this method.
virtual Status DeleteDir(const std::string& dirname);
// Store the size of fname in *file_size. // Store the size of fname in *file_size.
virtual Status GetFileSize(const std::string& fname, uint64_t* file_size) = 0; virtual Status GetFileSize(const std::string& fname, uint64_t* file_size) = 0;
@ -169,7 +110,7 @@ class LEVELDB_EXPORT Env {
const std::string& target) = 0; const std::string& target) = 0;
// Lock the specified file. Used to prevent concurrent access to // Lock the specified file. Used to prevent concurrent access to
// the same db by multiple processes. On failure, stores nullptr in // the same db by multiple processes. On failure, stores NULL in
// *lock and returns non-OK. // *lock and returns non-OK.
// //
// On success, stores a pointer to the object that represents the // On success, stores a pointer to the object that represents the
@ -195,14 +136,16 @@ class LEVELDB_EXPORT Env {
// added to the same Env may run concurrently in different threads. // added to the same Env may run concurrently in different threads.
// I.e., the caller may not assume that background work items are // I.e., the caller may not assume that background work items are
// serialized. // serialized.
virtual void Schedule(void (*function)(void* arg), void* arg) = 0; virtual void Schedule(
void (*function)(void* arg),
void* arg) = 0;
// Start a new thread, invoking "function(arg)" within the new thread. // Start a new thread, invoking "function(arg)" within the new thread.
// When "function(arg)" returns, the thread will be destroyed. // When "function(arg)" returns, the thread will be destroyed.
virtual void StartThread(void (*function)(void* arg), void* arg) = 0; virtual void StartThread(void (*function)(void* arg), void* arg) = 0;
// *path is set to a temporary directory that can be used for testing. It may // *path is set to a temporary directory that can be used for testing. It may
// or may not have just been created. The directory may or may not differ // or many not have just been created. The directory may or may not differ
// between runs of the same process, but subsequent calls will return the // between runs of the same process, but subsequent calls will return the
// same directory. // same directory.
virtual Status GetTestDirectory(std::string* path) = 0; virtual Status GetTestDirectory(std::string* path) = 0;
@ -216,16 +159,17 @@ class LEVELDB_EXPORT Env {
// Sleep/delay the thread for the prescribed number of micro-seconds. // Sleep/delay the thread for the prescribed number of micro-seconds.
virtual void SleepForMicroseconds(int micros) = 0; virtual void SleepForMicroseconds(int micros) = 0;
private:
// No copying allowed
Env(const Env&);
void operator=(const Env&);
}; };
// A file abstraction for reading sequentially through a file // A file abstraction for reading sequentially through a file
class LEVELDB_EXPORT SequentialFile { class SequentialFile {
public: public:
SequentialFile() = default; SequentialFile() { }
SequentialFile(const SequentialFile&) = delete;
SequentialFile& operator=(const SequentialFile&) = delete;
virtual ~SequentialFile(); virtual ~SequentialFile();
// Read up to "n" bytes from the file. "scratch[0..n-1]" may be // Read up to "n" bytes from the file. "scratch[0..n-1]" may be
@ -246,16 +190,17 @@ class LEVELDB_EXPORT SequentialFile {
// //
// REQUIRES: External synchronization // REQUIRES: External synchronization
virtual Status Skip(uint64_t n) = 0; virtual Status Skip(uint64_t n) = 0;
private:
// No copying allowed
SequentialFile(const SequentialFile&);
void operator=(const SequentialFile&);
}; };
// A file abstraction for randomly reading the contents of a file. // A file abstraction for randomly reading the contents of a file.
class LEVELDB_EXPORT RandomAccessFile { class RandomAccessFile {
public: public:
RandomAccessFile() = default; RandomAccessFile() { }
RandomAccessFile(const RandomAccessFile&) = delete;
RandomAccessFile& operator=(const RandomAccessFile&) = delete;
virtual ~RandomAccessFile(); virtual ~RandomAccessFile();
// Read up to "n" bytes from the file starting at "offset". // Read up to "n" bytes from the file starting at "offset".
@ -269,149 +214,138 @@ class LEVELDB_EXPORT RandomAccessFile {
// Safe for concurrent use by multiple threads. // Safe for concurrent use by multiple threads.
virtual Status Read(uint64_t offset, size_t n, Slice* result, virtual Status Read(uint64_t offset, size_t n, Slice* result,
char* scratch) const = 0; char* scratch) const = 0;
private:
// No copying allowed
RandomAccessFile(const RandomAccessFile&);
void operator=(const RandomAccessFile&);
}; };
// A file abstraction for sequential writing. The implementation // A file abstraction for sequential writing. The implementation
// must provide buffering since callers may append small fragments // must provide buffering since callers may append small fragments
// at a time to the file. // at a time to the file.
class LEVELDB_EXPORT WritableFile { class WritableFile {
public: public:
WritableFile() = default; WritableFile() { }
WritableFile(const WritableFile&) = delete;
WritableFile& operator=(const WritableFile&) = delete;
virtual ~WritableFile(); virtual ~WritableFile();
virtual Status Append(const Slice& data) = 0; virtual Status Append(const Slice& data) = 0;
virtual Status Close() = 0; virtual Status Close() = 0;
virtual Status Flush() = 0; virtual Status Flush() = 0;
virtual Status Sync() = 0; virtual Status Sync() = 0;
private:
// No copying allowed
WritableFile(const WritableFile&);
void operator=(const WritableFile&);
}; };
// An interface for writing log messages. // An interface for writing log messages.
class LEVELDB_EXPORT Logger { class Logger {
public: public:
Logger() = default; Logger() { }
Logger(const Logger&) = delete;
Logger& operator=(const Logger&) = delete;
virtual ~Logger(); virtual ~Logger();
// Write an entry to the log file with the specified format. // Write an entry to the log file with the specified format.
virtual void Logv(const char* format, std::va_list ap) = 0; virtual void Logv(const char* format, va_list ap) = 0;
private:
// No copying allowed
Logger(const Logger&);
void operator=(const Logger&);
}; };
// Identifies a locked file. // Identifies a locked file.
class LEVELDB_EXPORT FileLock { class FileLock {
public: public:
FileLock() = default; FileLock() { }
FileLock(const FileLock&) = delete;
FileLock& operator=(const FileLock&) = delete;
virtual ~FileLock(); virtual ~FileLock();
private:
// No copying allowed
FileLock(const FileLock&);
void operator=(const FileLock&);
}; };
// Log the specified data to *info_log if info_log is non-null. // Log the specified data to *info_log if info_log is non-NULL.
void Log(Logger* info_log, const char* format, ...) extern void Log(Logger* info_log, const char* format, ...)
#if defined(__GNUC__) || defined(__clang__) # if defined(__GNUC__) || defined(__clang__)
__attribute__((__format__(__printf__, 2, 3))) __attribute__((__format__ (__printf__, 2, 3)))
#endif # endif
; ;
// A utility routine: write "data" to the named file. // A utility routine: write "data" to the named file.
LEVELDB_EXPORT Status WriteStringToFile(Env* env, const Slice& data, extern Status WriteStringToFile(Env* env, const Slice& data,
const std::string& fname); const std::string& fname);
// A utility routine: read contents of named file into *data // A utility routine: read contents of named file into *data
LEVELDB_EXPORT Status ReadFileToString(Env* env, const std::string& fname, extern Status ReadFileToString(Env* env, const std::string& fname,
std::string* data); std::string* data);
// An implementation of Env that forwards all calls to another Env. // An implementation of Env that forwards all calls to another Env.
// May be useful to clients who wish to override just part of the // May be useful to clients who wish to override just part of the
// functionality of another Env. // functionality of another Env.
class LEVELDB_EXPORT EnvWrapper : public Env { class EnvWrapper : public Env {
public: public:
// Initialize an EnvWrapper that delegates all calls to *t. // Initialize an EnvWrapper that delegates all calls to *t
explicit EnvWrapper(Env* t) : target_(t) {} explicit EnvWrapper(Env* t) : target_(t) { }
virtual ~EnvWrapper(); virtual ~EnvWrapper();
// Return the target to which this Env forwards all calls. // Return the target to which this Env forwards all calls
Env* target() const { return target_; } Env* target() const { return target_; }
// The following text is boilerplate that forwards all methods to target(). // The following text is boilerplate that forwards all methods to target()
Status NewSequentialFile(const std::string& f, SequentialFile** r) override { Status NewSequentialFile(const std::string& f, SequentialFile** r) {
return target_->NewSequentialFile(f, r); return target_->NewSequentialFile(f, r);
} }
Status NewRandomAccessFile(const std::string& f, Status NewRandomAccessFile(const std::string& f, RandomAccessFile** r) {
RandomAccessFile** r) override {
return target_->NewRandomAccessFile(f, r); return target_->NewRandomAccessFile(f, r);
} }
Status NewWritableFile(const std::string& f, WritableFile** r) override { Status NewWritableFile(const std::string& f, WritableFile** r) {
return target_->NewWritableFile(f, r); return target_->NewWritableFile(f, r);
} }
Status NewAppendableFile(const std::string& f, WritableFile** r) override { Status NewAppendableFile(const std::string& f, WritableFile** r) {
return target_->NewAppendableFile(f, r); return target_->NewAppendableFile(f, r);
} }
bool FileExists(const std::string& f) override { bool FileExists(const std::string& f) { return target_->FileExists(f); }
return target_->FileExists(f); Status GetChildren(const std::string& dir, std::vector<std::string>* r) {
}
Status GetChildren(const std::string& dir,
std::vector<std::string>* r) override {
return target_->GetChildren(dir, r); return target_->GetChildren(dir, r);
} }
Status RemoveFile(const std::string& f) override { Status DeleteFile(const std::string& f) { return target_->DeleteFile(f); }
return target_->RemoveFile(f); Status CreateDir(const std::string& d) { return target_->CreateDir(d); }
} Status DeleteDir(const std::string& d) { return target_->DeleteDir(d); }
Status CreateDir(const std::string& d) override { Status GetFileSize(const std::string& f, uint64_t* s) {
return target_->CreateDir(d);
}
Status RemoveDir(const std::string& d) override {
return target_->RemoveDir(d);
}
Status GetFileSize(const std::string& f, uint64_t* s) override {
return target_->GetFileSize(f, s); return target_->GetFileSize(f, s);
} }
Status RenameFile(const std::string& s, const std::string& t) override { Status RenameFile(const std::string& s, const std::string& t) {
return target_->RenameFile(s, t); return target_->RenameFile(s, t);
} }
Status LockFile(const std::string& f, FileLock** l) override { Status LockFile(const std::string& f, FileLock** l) {
return target_->LockFile(f, l); return target_->LockFile(f, l);
} }
Status UnlockFile(FileLock* l) override { return target_->UnlockFile(l); } Status UnlockFile(FileLock* l) { return target_->UnlockFile(l); }
void Schedule(void (*f)(void*), void* a) override { void Schedule(void (*f)(void*), void* a) {
return target_->Schedule(f, a); return target_->Schedule(f, a);
} }
void StartThread(void (*f)(void*), void* a) override { void StartThread(void (*f)(void*), void* a) {
return target_->StartThread(f, a); return target_->StartThread(f, a);
} }
Status GetTestDirectory(std::string* path) override { virtual Status GetTestDirectory(std::string* path) {
return target_->GetTestDirectory(path); return target_->GetTestDirectory(path);
} }
Status NewLogger(const std::string& fname, Logger** result) override { virtual Status NewLogger(const std::string& fname, Logger** result) {
return target_->NewLogger(fname, result); return target_->NewLogger(fname, result);
} }
uint64_t NowMicros() override { return target_->NowMicros(); } uint64_t NowMicros() {
void SleepForMicroseconds(int micros) override { return target_->NowMicros();
}
void SleepForMicroseconds(int micros) {
target_->SleepForMicroseconds(micros); target_->SleepForMicroseconds(micros);
} }
private: private:
Env* target_; Env* target_;
}; };
} // namespace leveldb } // namespace leveldb
// This workaround can be removed when leveldb::Env::DeleteFile is removed.
// Redefine DeleteFile if it was undefined earlier.
#if defined(_WIN32) && defined(LEVELDB_DELETEFILE_UNDEFINED)
#if defined(UNICODE)
#define DeleteFile DeleteFileW
#else
#define DeleteFile DeleteFileA
#endif // defined(UNICODE)
#endif // defined(_WIN32) && defined(LEVELDB_DELETEFILE_UNDEFINED)
#endif // STORAGE_LEVELDB_INCLUDE_ENV_H_ #endif // STORAGE_LEVELDB_INCLUDE_ENV_H_

View File

@ -1,33 +0,0 @@
// Copyright (c) 2017 The LevelDB Authors. All rights reserved.
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file. See the AUTHORS file for names of contributors.
#ifndef STORAGE_LEVELDB_INCLUDE_EXPORT_H_
#define STORAGE_LEVELDB_INCLUDE_EXPORT_H_
#if !defined(LEVELDB_EXPORT)
#if defined(LEVELDB_SHARED_LIBRARY)
#if defined(_WIN32)
#if defined(LEVELDB_COMPILE_LIBRARY)
#define LEVELDB_EXPORT __declspec(dllexport)
#else
#define LEVELDB_EXPORT __declspec(dllimport)
#endif // defined(LEVELDB_COMPILE_LIBRARY)
#else // defined(_WIN32)
#if defined(LEVELDB_COMPILE_LIBRARY)
#define LEVELDB_EXPORT __attribute__((visibility("default")))
#else
#define LEVELDB_EXPORT
#endif
#endif // defined(_WIN32)
#else // defined(LEVELDB_SHARED_LIBRARY)
#define LEVELDB_EXPORT
#endif
#endif // !defined(LEVELDB_EXPORT)
#endif // STORAGE_LEVELDB_INCLUDE_EXPORT_H_

View File

@ -18,13 +18,11 @@
#include <string> #include <string>
#include "leveldb/export.h"
namespace leveldb { namespace leveldb {
class Slice; class Slice;
class LEVELDB_EXPORT FilterPolicy { class FilterPolicy {
public: public:
virtual ~FilterPolicy(); virtual ~FilterPolicy();
@ -40,8 +38,8 @@ class LEVELDB_EXPORT FilterPolicy {
// //
// Warning: do not change the initial contents of *dst. Instead, // Warning: do not change the initial contents of *dst. Instead,
// append the newly constructed filter to *dst. // append the newly constructed filter to *dst.
virtual void CreateFilter(const Slice* keys, int n, virtual void CreateFilter(const Slice* keys, int n, std::string* dst)
std::string* dst) const = 0; const = 0;
// "filter" contains the data appended by a preceding call to // "filter" contains the data appended by a preceding call to
// CreateFilter() on this class. This method must return true if // CreateFilter() on this class. This method must return true if
@ -65,8 +63,8 @@ class LEVELDB_EXPORT FilterPolicy {
// ignores trailing spaces, it would be incorrect to use a // ignores trailing spaces, it would be incorrect to use a
// FilterPolicy (like NewBloomFilterPolicy) that does not ignore // FilterPolicy (like NewBloomFilterPolicy) that does not ignore
// trailing spaces in keys. // trailing spaces in keys.
LEVELDB_EXPORT const FilterPolicy* NewBloomFilterPolicy(int bits_per_key); extern const FilterPolicy* NewBloomFilterPolicy(int bits_per_key);
} // namespace leveldb }
#endif // STORAGE_LEVELDB_INCLUDE_FILTER_POLICY_H_ #endif // STORAGE_LEVELDB_INCLUDE_FILTER_POLICY_H_

View File

@ -15,19 +15,14 @@
#ifndef STORAGE_LEVELDB_INCLUDE_ITERATOR_H_ #ifndef STORAGE_LEVELDB_INCLUDE_ITERATOR_H_
#define STORAGE_LEVELDB_INCLUDE_ITERATOR_H_ #define STORAGE_LEVELDB_INCLUDE_ITERATOR_H_
#include "leveldb/export.h"
#include "leveldb/slice.h" #include "leveldb/slice.h"
#include "leveldb/status.h" #include "leveldb/status.h"
namespace leveldb { namespace leveldb {
class LEVELDB_EXPORT Iterator { class Iterator {
public: public:
Iterator(); Iterator();
Iterator(const Iterator&) = delete;
Iterator& operator=(const Iterator&) = delete;
virtual ~Iterator(); virtual ~Iterator();
// An iterator is either positioned at a key/value pair, or // An iterator is either positioned at a key/value pair, or
@ -77,35 +72,28 @@ class LEVELDB_EXPORT Iterator {
// //
// Note that unlike all of the preceding methods, this method is // Note that unlike all of the preceding methods, this method is
// not abstract and therefore clients should not override it. // not abstract and therefore clients should not override it.
using CleanupFunction = void (*)(void* arg1, void* arg2); typedef void (*CleanupFunction)(void* arg1, void* arg2);
void RegisterCleanup(CleanupFunction function, void* arg1, void* arg2); void RegisterCleanup(CleanupFunction function, void* arg1, void* arg2);
private: private:
// Cleanup functions are stored in a single-linked list. struct Cleanup {
// The list's head node is inlined in the iterator.
struct CleanupNode {
// True if the node is not used. Only head nodes might be unused.
bool IsEmpty() const { return function == nullptr; }
// Invokes the cleanup function.
void Run() {
assert(function != nullptr);
(*function)(arg1, arg2);
}
// The head node is used if the function pointer is not null.
CleanupFunction function; CleanupFunction function;
void* arg1; void* arg1;
void* arg2; void* arg2;
CleanupNode* next; Cleanup* next;
}; };
CleanupNode cleanup_head_; Cleanup cleanup_;
// No copying allowed
Iterator(const Iterator&);
void operator=(const Iterator&);
}; };
// Return an empty iterator (yields nothing). // Return an empty iterator (yields nothing).
LEVELDB_EXPORT Iterator* NewEmptyIterator(); extern Iterator* NewEmptyIterator();
// Return an empty iterator with the specified status. // Return an empty iterator with the specified status.
LEVELDB_EXPORT Iterator* NewErrorIterator(const Status& status); extern Iterator* NewErrorIterator(const Status& status);
} // namespace leveldb } // namespace leveldb

View File

@ -5,9 +5,7 @@
#ifndef STORAGE_LEVELDB_INCLUDE_OPTIONS_H_ #ifndef STORAGE_LEVELDB_INCLUDE_OPTIONS_H_
#define STORAGE_LEVELDB_INCLUDE_OPTIONS_H_ #define STORAGE_LEVELDB_INCLUDE_OPTIONS_H_
#include <cstddef> #include <stddef.h>
#include "leveldb/export.h"
namespace leveldb { namespace leveldb {
@ -25,16 +23,12 @@ class Snapshot;
enum CompressionType { enum CompressionType {
// NOTE: do not change the values of existing entries, as these are // NOTE: do not change the values of existing entries, as these are
// part of the persistent format on disk. // part of the persistent format on disk.
kNoCompression = 0x0, kNoCompression = 0x0,
kSnappyCompression = 0x1, kSnappyCompression = 0x1
kZstdCompression = 0x2,
}; };
// Options to control the behavior of a database (passed to DB::Open) // Options to control the behavior of a database (passed to DB::Open)
struct LEVELDB_EXPORT Options { struct Options {
// Create an Options object with default values for all fields.
Options();
// ------------------- // -------------------
// Parameters that affect behavior // Parameters that affect behavior
@ -47,17 +41,20 @@ struct LEVELDB_EXPORT Options {
const Comparator* comparator; const Comparator* comparator;
// If true, the database will be created if it is missing. // If true, the database will be created if it is missing.
bool create_if_missing = false; // Default: false
bool create_if_missing;
// If true, an error is raised if the database already exists. // If true, an error is raised if the database already exists.
bool error_if_exists = false; // Default: false
bool error_if_exists;
// If true, the implementation will do aggressive checking of the // If true, the implementation will do aggressive checking of the
// data it is processing and will stop early if it detects any // data it is processing and will stop early if it detects any
// errors. This may have unforeseen ramifications: for example, a // errors. This may have unforeseen ramifications: for example, a
// corruption of one DB entry may cause a large number of entries to // corruption of one DB entry may cause a large number of entries to
// become unreadable or for the entire DB to become unopenable. // become unreadable or for the entire DB to become unopenable.
bool paranoid_checks = false; // Default: false
bool paranoid_checks;
// Use the specified object to interact with the environment, // Use the specified object to interact with the environment,
// e.g. to read/write files, schedule background work, etc. // e.g. to read/write files, schedule background work, etc.
@ -65,9 +62,10 @@ struct LEVELDB_EXPORT Options {
Env* env; Env* env;
// Any internal progress/error information generated by the db will // Any internal progress/error information generated by the db will
// be written to info_log if it is non-null, or to a file stored // be written to info_log if it is non-NULL, or to a file stored
// in the same directory as the DB contents if info_log is null. // in the same directory as the DB contents if info_log is NULL.
Logger* info_log = nullptr; // Default: NULL
Logger* info_log;
// ------------------- // -------------------
// Parameters that affect performance // Parameters that affect performance
@ -80,40 +78,39 @@ struct LEVELDB_EXPORT Options {
// so you may wish to adjust this parameter to control memory usage. // so you may wish to adjust this parameter to control memory usage.
// Also, a larger write buffer will result in a longer recovery time // Also, a larger write buffer will result in a longer recovery time
// the next time the database is opened. // the next time the database is opened.
size_t write_buffer_size = 4 * 1024 * 1024; //
// Default: 4MB
size_t write_buffer_size;
// Number of open files that can be used by the DB. You may need to // Number of open files that can be used by the DB. You may need to
// increase this if your database has a large working set (budget // increase this if your database has a large working set (budget
// one open file per 2MB of working set). // one open file per 2MB of working set).
int max_open_files = 1000; //
// Default: 1000
int max_open_files;
// Control over blocks (user data is stored in a set of blocks, and // Control over blocks (user data is stored in a set of blocks, and
// a block is the unit of reading from disk). // a block is the unit of reading from disk).
// If non-null, use the specified cache for blocks. // If non-NULL, use the specified cache for blocks.
// If null, leveldb will automatically create and use an 8MB internal cache. // If NULL, leveldb will automatically create and use an 8MB internal cache.
Cache* block_cache = nullptr; // Default: NULL
Cache* block_cache;
// Approximate size of user data packed per block. Note that the // Approximate size of user data packed per block. Note that the
// block size specified here corresponds to uncompressed data. The // block size specified here corresponds to uncompressed data. The
// actual size of the unit read from disk may be smaller if // actual size of the unit read from disk may be smaller if
// compression is enabled. This parameter can be changed dynamically. // compression is enabled. This parameter can be changed dynamically.
size_t block_size = 4 * 1024; //
// Default: 4K
size_t block_size;
// Number of keys between restart points for delta encoding of keys. // Number of keys between restart points for delta encoding of keys.
// This parameter can be changed dynamically. Most clients should // This parameter can be changed dynamically. Most clients should
// leave this parameter alone. // leave this parameter alone.
int block_restart_interval = 16; //
// Default: 16
// Leveldb will write up to this amount of bytes to a file before int block_restart_interval;
// switching to a new one.
// Most clients should leave this parameter alone. However if your
// filesystem is more efficient with larger files, you could
// consider increasing the value. The downside will be longer
// compactions and hence longer latency/performance hiccups.
// Another reason to increase this parameter might be when you are
// initially populating a large database.
size_t max_file_size = 2 * 1024 * 1024;
// Compress blocks using the specified compression algorithm. This // Compress blocks using the specified compression algorithm. This
// parameter can be changed dynamically. // parameter can be changed dynamically.
@ -129,45 +126,53 @@ struct LEVELDB_EXPORT Options {
// worth switching to kNoCompression. Even if the input data is // worth switching to kNoCompression. Even if the input data is
// incompressible, the kSnappyCompression implementation will // incompressible, the kSnappyCompression implementation will
// efficiently detect that and will switch to uncompressed mode. // efficiently detect that and will switch to uncompressed mode.
CompressionType compression = kSnappyCompression; CompressionType compression;
// Compression level for zstd.
// Currently only the range [-5,22] is supported. Default is 1.
int zstd_compression_level = 1;
// EXPERIMENTAL: If true, append to existing MANIFEST and log files // EXPERIMENTAL: If true, append to existing MANIFEST and log files
// when a database is opened. This can significantly speed up open. // when a database is opened. This can significantly speed up open.
// //
// Default: currently false, but may become true later. // Default: currently false, but may become true later.
bool reuse_logs = false; bool reuse_logs;
// If non-null, use the specified filter policy to reduce disk reads. // If non-NULL, use the specified filter policy to reduce disk reads.
// Many applications will benefit from passing the result of // Many applications will benefit from passing the result of
// NewBloomFilterPolicy() here. // NewBloomFilterPolicy() here.
const FilterPolicy* filter_policy = nullptr; //
// Default: NULL
const FilterPolicy* filter_policy;
// Create an Options object with default values for all fields.
Options();
}; };
// Options that control read operations // Options that control read operations
struct LEVELDB_EXPORT ReadOptions { struct ReadOptions {
// If true, all data read from underlying storage will be // If true, all data read from underlying storage will be
// verified against corresponding checksums. // verified against corresponding checksums.
bool verify_checksums = false; // Default: false
bool verify_checksums;
// Should the data read for this iteration be cached in memory? // Should the data read for this iteration be cached in memory?
// Callers may wish to set this field to false for bulk scans. // Callers may wish to set this field to false for bulk scans.
bool fill_cache = true; // Default: true
bool fill_cache;
// If "snapshot" is non-null, read as of the supplied snapshot // If "snapshot" is non-NULL, read as of the supplied snapshot
// (which must belong to the DB that is being read and which must // (which must belong to the DB that is being read and which must
// not have been released). If "snapshot" is null, use an implicit // not have been released). If "snapshot" is NULL, use an implicit
// snapshot of the state at the beginning of this read operation. // snapshot of the state at the beginning of this read operation.
const Snapshot* snapshot = nullptr; // Default: NULL
const Snapshot* snapshot;
ReadOptions()
: verify_checksums(false),
fill_cache(true),
snapshot(NULL) {
}
}; };
// Options that control write operations // Options that control write operations
struct LEVELDB_EXPORT WriteOptions { struct WriteOptions {
WriteOptions() = default;
// If true, the write will be flushed from the operating system // If true, the write will be flushed from the operating system
// buffer cache (by calling WritableFile::Sync()) before the write // buffer cache (by calling WritableFile::Sync()) before the write
// is considered complete. If this flag is true, writes will be // is considered complete. If this flag is true, writes will be
@ -182,7 +187,13 @@ struct LEVELDB_EXPORT WriteOptions {
// crash semantics as the "write()" system call. A DB write // crash semantics as the "write()" system call. A DB write
// with sync==true has similar crash semantics to a "write()" // with sync==true has similar crash semantics to a "write()"
// system call followed by "fsync()". // system call followed by "fsync()".
bool sync = false; //
// Default: false
bool sync;
WriteOptions()
: sync(false) {
}
}; };
} // namespace leveldb } // namespace leveldb

View File

@ -15,32 +15,26 @@
#ifndef STORAGE_LEVELDB_INCLUDE_SLICE_H_ #ifndef STORAGE_LEVELDB_INCLUDE_SLICE_H_
#define STORAGE_LEVELDB_INCLUDE_SLICE_H_ #define STORAGE_LEVELDB_INCLUDE_SLICE_H_
#include <cassert> #include <assert.h>
#include <cstddef> #include <stddef.h>
#include <cstring> #include <string.h>
#include <string> #include <string>
#include "leveldb/export.h"
namespace leveldb { namespace leveldb {
class LEVELDB_EXPORT Slice { class Slice {
public: public:
// Create an empty slice. // Create an empty slice.
Slice() : data_(""), size_(0) {} Slice() : data_(""), size_(0) { }
// Create a slice that refers to d[0,n-1]. // Create a slice that refers to d[0,n-1].
Slice(const char* d, size_t n) : data_(d), size_(n) {} Slice(const char* d, size_t n) : data_(d), size_(n) { }
// Create a slice that refers to the contents of "s" // Create a slice that refers to the contents of "s"
Slice(const std::string& s) : data_(s.data()), size_(s.size()) {} Slice(const std::string& s) : data_(s.data()), size_(s.size()) { }
// Create a slice that refers to s[0,strlen(s)-1] // Create a slice that refers to s[0,strlen(s)-1]
Slice(const char* s) : data_(s), size_(strlen(s)) {} Slice(const char* s) : data_(s), size_(strlen(s)) { }
// Intentionally copyable.
Slice(const Slice&) = default;
Slice& operator=(const Slice&) = default;
// Return a pointer to the beginning of the referenced data // Return a pointer to the beginning of the referenced data
const char* data() const { return data_; } const char* data() const { return data_; }
@ -59,10 +53,7 @@ class LEVELDB_EXPORT Slice {
} }
// Change this slice to refer to an empty array // Change this slice to refer to an empty array
void clear() { void clear() { data_ = ""; size_ = 0; }
data_ = "";
size_ = 0;
}
// Drop the first "n" bytes from this slice. // Drop the first "n" bytes from this slice.
void remove_prefix(size_t n) { void remove_prefix(size_t n) {
@ -82,12 +73,15 @@ class LEVELDB_EXPORT Slice {
// Return true iff "x" is a prefix of "*this" // Return true iff "x" is a prefix of "*this"
bool starts_with(const Slice& x) const { bool starts_with(const Slice& x) const {
return ((size_ >= x.size_) && (memcmp(data_, x.data_, x.size_) == 0)); return ((size_ >= x.size_) &&
(memcmp(data_, x.data_, x.size_) == 0));
} }
private: private:
const char* data_; const char* data_;
size_t size_; size_t size_;
// Intentionally copyable
}; };
inline bool operator==(const Slice& x, const Slice& y) { inline bool operator==(const Slice& x, const Slice& y) {
@ -95,20 +89,21 @@ inline bool operator==(const Slice& x, const Slice& y) {
(memcmp(x.data(), y.data(), x.size()) == 0)); (memcmp(x.data(), y.data(), x.size()) == 0));
} }
inline bool operator!=(const Slice& x, const Slice& y) { return !(x == y); } inline bool operator!=(const Slice& x, const Slice& y) {
return !(x == y);
}
inline int Slice::compare(const Slice& b) const { inline int Slice::compare(const Slice& b) const {
const size_t min_len = (size_ < b.size_) ? size_ : b.size_; const size_t min_len = (size_ < b.size_) ? size_ : b.size_;
int r = memcmp(data_, b.data_, min_len); int r = memcmp(data_, b.data_, min_len);
if (r == 0) { if (r == 0) {
if (size_ < b.size_) if (size_ < b.size_) r = -1;
r = -1; else if (size_ > b.size_) r = +1;
else if (size_ > b.size_)
r = +1;
} }
return r; return r;
} }
} // namespace leveldb } // namespace leveldb
#endif // STORAGE_LEVELDB_INCLUDE_SLICE_H_ #endif // STORAGE_LEVELDB_INCLUDE_SLICE_H_

View File

@ -13,25 +13,20 @@
#ifndef STORAGE_LEVELDB_INCLUDE_STATUS_H_ #ifndef STORAGE_LEVELDB_INCLUDE_STATUS_H_
#define STORAGE_LEVELDB_INCLUDE_STATUS_H_ #define STORAGE_LEVELDB_INCLUDE_STATUS_H_
#include <algorithm>
#include <string> #include <string>
#include "leveldb/export.h"
#include "leveldb/slice.h" #include "leveldb/slice.h"
namespace leveldb { namespace leveldb {
class LEVELDB_EXPORT Status { class Status {
public: public:
// Create a success status. // Create a success status.
Status() noexcept : state_(nullptr) {} Status() : state_(NULL) { }
~Status() { delete[] state_; } ~Status() { delete[] state_; }
Status(const Status& rhs); // Copy the specified status.
Status& operator=(const Status& rhs); Status(const Status& s);
void operator=(const Status& s);
Status(Status&& rhs) noexcept : state_(rhs.state_) { rhs.state_ = nullptr; }
Status& operator=(Status&& rhs) noexcept;
// Return a success status. // Return a success status.
static Status OK() { return Status(); } static Status OK() { return Status(); }
@ -54,7 +49,7 @@ class LEVELDB_EXPORT Status {
} }
// Returns true iff the status indicates success. // Returns true iff the status indicates success.
bool ok() const { return (state_ == nullptr); } bool ok() const { return (state_ == NULL); }
// Returns true iff the status indicates a NotFound error. // Returns true iff the status indicates a NotFound error.
bool IsNotFound() const { return code() == kNotFound; } bool IsNotFound() const { return code() == kNotFound; }
@ -68,14 +63,18 @@ class LEVELDB_EXPORT Status {
// Returns true iff the status indicates a NotSupportedError. // Returns true iff the status indicates a NotSupportedError.
bool IsNotSupportedError() const { return code() == kNotSupported; } bool IsNotSupportedError() const { return code() == kNotSupported; }
// Returns true iff the status indicates an InvalidArgument.
bool IsInvalidArgument() const { return code() == kInvalidArgument; }
// Return a string representation of this status suitable for printing. // Return a string representation of this status suitable for printing.
// Returns the string "OK" for success. // Returns the string "OK" for success.
std::string ToString() const; std::string ToString() const;
private: private:
// OK status has a NULL state_. Otherwise, state_ is a new[] array
// of the following form:
// state_[0..3] == length of message
// state_[4] == code
// state_[5..] == message
const char* state_;
enum Code { enum Code {
kOk = 0, kOk = 0,
kNotFound = 1, kNotFound = 1,
@ -86,35 +85,23 @@ class LEVELDB_EXPORT Status {
}; };
Code code() const { Code code() const {
return (state_ == nullptr) ? kOk : static_cast<Code>(state_[4]); return (state_ == NULL) ? kOk : static_cast<Code>(state_[4]);
} }
Status(Code code, const Slice& msg, const Slice& msg2); Status(Code code, const Slice& msg, const Slice& msg2);
static const char* CopyState(const char* s); static const char* CopyState(const char* s);
// OK status has a null state_. Otherwise, state_ is a new[] array
// of the following form:
// state_[0..3] == length of message
// state_[4] == code
// state_[5..] == message
const char* state_;
}; };
inline Status::Status(const Status& rhs) { inline Status::Status(const Status& s) {
state_ = (rhs.state_ == nullptr) ? nullptr : CopyState(rhs.state_); state_ = (s.state_ == NULL) ? NULL : CopyState(s.state_);
} }
inline Status& Status::operator=(const Status& rhs) { inline void Status::operator=(const Status& s) {
// The following condition catches both aliasing (when this == &rhs), // The following condition catches both aliasing (when this == &s),
// and the common case where both rhs and *this are ok. // and the common case where both s and *this are ok.
if (state_ != rhs.state_) { if (state_ != s.state_) {
delete[] state_; delete[] state_;
state_ = (rhs.state_ == nullptr) ? nullptr : CopyState(rhs.state_); state_ = (s.state_ == NULL) ? NULL : CopyState(s.state_);
} }
return *this;
}
inline Status& Status::operator=(Status&& rhs) noexcept {
std::swap(state_, rhs.state_);
return *this;
} }
} // namespace leveldb } // namespace leveldb

View File

@ -5,9 +5,7 @@
#ifndef STORAGE_LEVELDB_INCLUDE_TABLE_H_ #ifndef STORAGE_LEVELDB_INCLUDE_TABLE_H_
#define STORAGE_LEVELDB_INCLUDE_TABLE_H_ #define STORAGE_LEVELDB_INCLUDE_TABLE_H_
#include <cstdint> #include <stdint.h>
#include "leveldb/export.h"
#include "leveldb/iterator.h" #include "leveldb/iterator.h"
namespace leveldb { namespace leveldb {
@ -23,7 +21,7 @@ class TableCache;
// A Table is a sorted map from strings to strings. Tables are // A Table is a sorted map from strings to strings. Tables are
// immutable and persistent. A Table may be safely accessed from // immutable and persistent. A Table may be safely accessed from
// multiple threads without external synchronization. // multiple threads without external synchronization.
class LEVELDB_EXPORT Table { class Table {
public: public:
// Attempt to open the table that is stored in bytes [0..file_size) // Attempt to open the table that is stored in bytes [0..file_size)
// of "file", and read the metadata entries necessary to allow // of "file", and read the metadata entries necessary to allow
@ -32,16 +30,15 @@ class LEVELDB_EXPORT Table {
// If successful, returns ok and sets "*table" to the newly opened // If successful, returns ok and sets "*table" to the newly opened
// table. The client should delete "*table" when no longer needed. // table. The client should delete "*table" when no longer needed.
// If there was an error while initializing the table, sets "*table" // If there was an error while initializing the table, sets "*table"
// to nullptr and returns a non-ok status. Does not take ownership of // to NULL and returns a non-ok status. Does not take ownership of
// "*source", but the client must ensure that "source" remains live // "*source", but the client must ensure that "source" remains live
// for the duration of the returned table's lifetime. // for the duration of the returned table's lifetime.
// //
// *file must remain live while this Table is in use. // *file must remain live while this Table is in use.
static Status Open(const Options& options, RandomAccessFile* file, static Status Open(const Options& options,
uint64_t file_size, Table** table); RandomAccessFile* file,
uint64_t file_size,
Table(const Table&) = delete; Table** table);
Table& operator=(const Table&) = delete;
~Table(); ~Table();
@ -59,24 +56,28 @@ class LEVELDB_EXPORT Table {
uint64_t ApproximateOffsetOf(const Slice& key) const; uint64_t ApproximateOffsetOf(const Slice& key) const;
private: private:
friend class TableCache;
struct Rep; struct Rep;
Rep* rep_;
explicit Table(Rep* rep) { rep_ = rep; }
static Iterator* BlockReader(void*, const ReadOptions&, const Slice&); static Iterator* BlockReader(void*, const ReadOptions&, const Slice&);
explicit Table(Rep* rep) : rep_(rep) {}
// Calls (*handle_result)(arg, ...) with the entry found after a call // Calls (*handle_result)(arg, ...) with the entry found after a call
// to Seek(key). May not make such a call if filter policy says // to Seek(key). May not make such a call if filter policy says
// that key is not present. // that key is not present.
Status InternalGet(const ReadOptions&, const Slice& key, void* arg, friend class TableCache;
void (*handle_result)(void* arg, const Slice& k, Status InternalGet(
const Slice& v)); const ReadOptions&, const Slice& key,
void* arg,
void (*handle_result)(void* arg, const Slice& k, const Slice& v));
void ReadMeta(const Footer& footer); void ReadMeta(const Footer& footer);
void ReadFilter(const Slice& filter_handle_value); void ReadFilter(const Slice& filter_handle_value);
Rep* const rep_; // No copying allowed
Table(const Table&);
void operator=(const Table&);
}; };
} // namespace leveldb } // namespace leveldb

View File

@ -13,9 +13,7 @@
#ifndef STORAGE_LEVELDB_INCLUDE_TABLE_BUILDER_H_ #ifndef STORAGE_LEVELDB_INCLUDE_TABLE_BUILDER_H_
#define STORAGE_LEVELDB_INCLUDE_TABLE_BUILDER_H_ #define STORAGE_LEVELDB_INCLUDE_TABLE_BUILDER_H_
#include <cstdint> #include <stdint.h>
#include "leveldb/export.h"
#include "leveldb/options.h" #include "leveldb/options.h"
#include "leveldb/status.h" #include "leveldb/status.h"
@ -25,16 +23,13 @@ class BlockBuilder;
class BlockHandle; class BlockHandle;
class WritableFile; class WritableFile;
class LEVELDB_EXPORT TableBuilder { class TableBuilder {
public: public:
// Create a builder that will store the contents of the table it is // Create a builder that will store the contents of the table it is
// building in *file. Does not close the file. It is up to the // building in *file. Does not close the file. It is up to the
// caller to close the file after calling Finish(). // caller to close the file after calling Finish().
TableBuilder(const Options& options, WritableFile* file); TableBuilder(const Options& options, WritableFile* file);
TableBuilder(const TableBuilder&) = delete;
TableBuilder& operator=(const TableBuilder&) = delete;
// REQUIRES: Either Finish() or Abandon() has been called. // REQUIRES: Either Finish() or Abandon() has been called.
~TableBuilder(); ~TableBuilder();
@ -86,6 +81,10 @@ class LEVELDB_EXPORT TableBuilder {
struct Rep; struct Rep;
Rep* rep_; Rep* rep_;
// No copying allowed
TableBuilder(const TableBuilder&);
void operator=(const TableBuilder&);
}; };
} // namespace leveldb } // namespace leveldb

View File

@ -22,29 +22,15 @@
#define STORAGE_LEVELDB_INCLUDE_WRITE_BATCH_H_ #define STORAGE_LEVELDB_INCLUDE_WRITE_BATCH_H_
#include <string> #include <string>
#include "leveldb/export.h"
#include "leveldb/status.h" #include "leveldb/status.h"
namespace leveldb { namespace leveldb {
class Slice; class Slice;
class LEVELDB_EXPORT WriteBatch { class WriteBatch {
public: public:
class LEVELDB_EXPORT Handler {
public:
virtual ~Handler();
virtual void Put(const Slice& key, const Slice& value) = 0;
virtual void Delete(const Slice& key) = 0;
};
WriteBatch(); WriteBatch();
// Intentionally copyable.
WriteBatch(const WriteBatch&) = default;
WriteBatch& operator=(const WriteBatch&) = default;
~WriteBatch(); ~WriteBatch();
// Store the mapping "key->value" in the database. // Store the mapping "key->value" in the database.
@ -56,26 +42,21 @@ class LEVELDB_EXPORT WriteBatch {
// Clear all updates buffered in this batch. // Clear all updates buffered in this batch.
void Clear(); void Clear();
// The size of the database changes caused by this batch.
//
// This number is tied to implementation details, and may change across
// releases. It is intended for LevelDB usage metrics.
size_t ApproximateSize() const;
// Copies the operations in "source" to this batch.
//
// This runs in O(source size) time. However, the constant factor is better
// than calling Iterate() over the source batch with a Handler that replicates
// the operations into this batch.
void Append(const WriteBatch& source);
// Support for iterating over the contents of a batch. // Support for iterating over the contents of a batch.
class Handler {
public:
virtual ~Handler();
virtual void Put(const Slice& key, const Slice& value) = 0;
virtual void Delete(const Slice& key) = 0;
};
Status Iterate(Handler* handler) const; Status Iterate(Handler* handler) const;
private: private:
friend class WriteBatchInternal; friend class WriteBatchInternal;
std::string rep_; // See comment in write_batch.cc for the format of rep_ std::string rep_; // See comment in write_batch.cc for the format of rep_
// Intentionally copyable
}; };
} // namespace leveldb } // namespace leveldb

View File

@ -3,14 +3,13 @@
// found in the LICENSE file. See the AUTHORS file for names of contributors. // found in the LICENSE file. See the AUTHORS file for names of contributors.
// Test for issue 178: a manual compaction causes deleted data to reappear. // Test for issue 178: a manual compaction causes deleted data to reappear.
#include <cstdlib>
#include <iostream> #include <iostream>
#include <sstream> #include <sstream>
#include <cstdlib>
#include "gtest/gtest.h"
#include "leveldb/db.h" #include "leveldb/db.h"
#include "leveldb/write_batch.h" #include "leveldb/write_batch.h"
#include "util/testutil.h" #include "util/testharness.h"
namespace { namespace {
@ -18,15 +17,19 @@ const int kNumKeys = 1100000;
std::string Key1(int i) { std::string Key1(int i) {
char buf[100]; char buf[100];
std::snprintf(buf, sizeof(buf), "my_key_%d", i); snprintf(buf, sizeof(buf), "my_key_%d", i);
return buf; return buf;
} }
std::string Key2(int i) { return Key1(i) + "_xxx"; } std::string Key2(int i) {
return Key1(i) + "_xxx";
}
class Issue178 { };
TEST(Issue178, Test) { TEST(Issue178, Test) {
// Get rid of any state from an old run. // Get rid of any state from an old run.
std::string dbpath = testing::TempDir() + "leveldb_cbug_test"; std::string dbpath = leveldb::test::TmpDir() + "/leveldb_cbug_test";
DestroyDB(dbpath, leveldb::Options()); DestroyDB(dbpath, leveldb::Options());
// Open database. Disable compression since it affects the creation // Open database. Disable compression since it affects the creation
@ -36,28 +39,28 @@ TEST(Issue178, Test) {
leveldb::Options db_options; leveldb::Options db_options;
db_options.create_if_missing = true; db_options.create_if_missing = true;
db_options.compression = leveldb::kNoCompression; db_options.compression = leveldb::kNoCompression;
ASSERT_LEVELDB_OK(leveldb::DB::Open(db_options, dbpath, &db)); ASSERT_OK(leveldb::DB::Open(db_options, dbpath, &db));
// create first key range // create first key range
leveldb::WriteBatch batch; leveldb::WriteBatch batch;
for (size_t i = 0; i < kNumKeys; i++) { for (size_t i = 0; i < kNumKeys; i++) {
batch.Put(Key1(i), "value for range 1 key"); batch.Put(Key1(i), "value for range 1 key");
} }
ASSERT_LEVELDB_OK(db->Write(leveldb::WriteOptions(), &batch)); ASSERT_OK(db->Write(leveldb::WriteOptions(), &batch));
// create second key range // create second key range
batch.Clear(); batch.Clear();
for (size_t i = 0; i < kNumKeys; i++) { for (size_t i = 0; i < kNumKeys; i++) {
batch.Put(Key2(i), "value for range 2 key"); batch.Put(Key2(i), "value for range 2 key");
} }
ASSERT_LEVELDB_OK(db->Write(leveldb::WriteOptions(), &batch)); ASSERT_OK(db->Write(leveldb::WriteOptions(), &batch));
// delete second key range // delete second key range
batch.Clear(); batch.Clear();
for (size_t i = 0; i < kNumKeys; i++) { for (size_t i = 0; i < kNumKeys; i++) {
batch.Delete(Key2(i)); batch.Delete(Key2(i));
} }
ASSERT_LEVELDB_OK(db->Write(leveldb::WriteOptions(), &batch)); ASSERT_OK(db->Write(leveldb::WriteOptions(), &batch));
// compact database // compact database
std::string start_key = Key1(0); std::string start_key = Key1(0);
@ -83,3 +86,7 @@ TEST(Issue178, Test) {
} }
} // anonymous namespace } // anonymous namespace
int main(int argc, char** argv) {
return leveldb::test::RunAllTests();
}

View File

@ -6,34 +6,35 @@
// to forward, the current key can be yielded unexpectedly if a new // to forward, the current key can be yielded unexpectedly if a new
// mutation has been added just before the current key. // mutation has been added just before the current key.
#include "gtest/gtest.h"
#include "leveldb/db.h" #include "leveldb/db.h"
#include "util/testutil.h" #include "util/testharness.h"
namespace leveldb { namespace leveldb {
class Issue200 { };
TEST(Issue200, Test) { TEST(Issue200, Test) {
// Get rid of any state from an old run. // Get rid of any state from an old run.
std::string dbpath = testing::TempDir() + "leveldb_issue200_test"; std::string dbpath = test::TmpDir() + "/leveldb_issue200_test";
DestroyDB(dbpath, Options()); DestroyDB(dbpath, Options());
DB* db; DB *db;
Options options; Options options;
options.create_if_missing = true; options.create_if_missing = true;
ASSERT_LEVELDB_OK(DB::Open(options, dbpath, &db)); ASSERT_OK(DB::Open(options, dbpath, &db));
WriteOptions write_options; WriteOptions write_options;
ASSERT_LEVELDB_OK(db->Put(write_options, "1", "b")); ASSERT_OK(db->Put(write_options, "1", "b"));
ASSERT_LEVELDB_OK(db->Put(write_options, "2", "c")); ASSERT_OK(db->Put(write_options, "2", "c"));
ASSERT_LEVELDB_OK(db->Put(write_options, "3", "d")); ASSERT_OK(db->Put(write_options, "3", "d"));
ASSERT_LEVELDB_OK(db->Put(write_options, "4", "e")); ASSERT_OK(db->Put(write_options, "4", "e"));
ASSERT_LEVELDB_OK(db->Put(write_options, "5", "f")); ASSERT_OK(db->Put(write_options, "5", "f"));
ReadOptions read_options; ReadOptions read_options;
Iterator* iter = db->NewIterator(read_options); Iterator *iter = db->NewIterator(read_options);
// Add an element that should not be reflected in the iterator. // Add an element that should not be reflected in the iterator.
ASSERT_LEVELDB_OK(db->Put(write_options, "25", "cd")); ASSERT_OK(db->Put(write_options, "25", "cd"));
iter->Seek("5"); iter->Seek("5");
ASSERT_EQ(iter->key().ToString(), "5"); ASSERT_EQ(iter->key().ToString(), "5");
@ -52,3 +53,7 @@ TEST(Issue200, Test) {
} }
} // namespace leveldb } // namespace leveldb
int main(int argc, char** argv) {
return leveldb::test::RunAllTests();
}

View File

@ -1,126 +0,0 @@
// Copyright (c) 2019 The LevelDB Authors. All rights reserved.
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file. See the AUTHORS file for names of contributors.
#include <cstdint>
#include <cstdlib>
#include <iostream>
#include <memory>
#include <string>
#include <vector>
#include "gtest/gtest.h"
#include "leveldb/db.h"
#include "leveldb/write_batch.h"
#include "util/testutil.h"
namespace leveldb {
namespace {
// Creates a random number in the range of [0, max).
int GenerateRandomNumber(int max) { return std::rand() % max; }
std::string CreateRandomString(int32_t index) {
static const size_t len = 1024;
char bytes[len];
size_t i = 0;
while (i < 8) {
bytes[i] = 'a' + ((index >> (4 * i)) & 0xf);
++i;
}
while (i < sizeof(bytes)) {
bytes[i] = 'a' + GenerateRandomNumber(26);
++i;
}
return std::string(bytes, sizeof(bytes));
}
} // namespace
TEST(Issue320, Test) {
std::srand(0);
bool delete_before_put = false;
bool keep_snapshots = true;
std::vector<std::unique_ptr<std::pair<std::string, std::string>>> test_map(
10000);
std::vector<Snapshot const*> snapshots(100, nullptr);
DB* db;
Options options;
options.create_if_missing = true;
std::string dbpath = testing::TempDir() + "leveldb_issue320_test";
ASSERT_LEVELDB_OK(DB::Open(options, dbpath, &db));
uint32_t target_size = 10000;
uint32_t num_items = 0;
uint32_t count = 0;
std::string key;
std::string value, old_value;
WriteOptions writeOptions;
ReadOptions readOptions;
while (count < 200000) {
if ((++count % 1000) == 0) {
std::cout << "count: " << count << std::endl;
}
int index = GenerateRandomNumber(test_map.size());
WriteBatch batch;
if (test_map[index] == nullptr) {
num_items++;
test_map[index].reset(new std::pair<std::string, std::string>(
CreateRandomString(index), CreateRandomString(index)));
batch.Put(test_map[index]->first, test_map[index]->second);
} else {
ASSERT_LEVELDB_OK(
db->Get(readOptions, test_map[index]->first, &old_value));
if (old_value != test_map[index]->second) {
std::cout << "ERROR incorrect value returned by Get" << std::endl;
std::cout << " count=" << count << std::endl;
std::cout << " old value=" << old_value << std::endl;
std::cout << " test_map[index]->second=" << test_map[index]->second
<< std::endl;
std::cout << " test_map[index]->first=" << test_map[index]->first
<< std::endl;
std::cout << " index=" << index << std::endl;
ASSERT_EQ(old_value, test_map[index]->second);
}
if (num_items >= target_size && GenerateRandomNumber(100) > 30) {
batch.Delete(test_map[index]->first);
test_map[index] = nullptr;
--num_items;
} else {
test_map[index]->second = CreateRandomString(index);
if (delete_before_put) batch.Delete(test_map[index]->first);
batch.Put(test_map[index]->first, test_map[index]->second);
}
}
ASSERT_LEVELDB_OK(db->Write(writeOptions, &batch));
if (keep_snapshots && GenerateRandomNumber(10) == 0) {
int i = GenerateRandomNumber(snapshots.size());
if (snapshots[i] != nullptr) {
db->ReleaseSnapshot(snapshots[i]);
}
snapshots[i] = db->GetSnapshot();
}
}
for (Snapshot const* snapshot : snapshots) {
if (snapshot) {
db->ReleaseSnapshot(snapshot);
}
}
delete db;
DestroyDB(dbpath, options);
}
} // namespace leveldb

View File

@ -5,6 +5,6 @@ Code in the rest of the package includes "port.h" from this directory.
"port.h" in turn includes a platform specific "port_<platform>.h" file "port.h" in turn includes a platform specific "port_<platform>.h" file
that provides the platform specific implementation. that provides the platform specific implementation.
See port_stdcxx.h for an example of what must be provided in a platform See port_posix.h for an example of what must be provided in a platform
specific header file. specific header file.

233
port/atomic_pointer.h Normal file
View File

@ -0,0 +1,233 @@
// Copyright (c) 2011 The LevelDB Authors. All rights reserved.
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file. See the AUTHORS file for names of contributors.
// AtomicPointer provides storage for a lock-free pointer.
// Platform-dependent implementation of AtomicPointer:
// - If the platform provides a cheap barrier, we use it with raw pointers
// - If <atomic> is present (on newer versions of gcc, it is), we use
// a <atomic>-based AtomicPointer. However we prefer the memory
// barrier based version, because at least on a gcc 4.4 32-bit build
// on linux, we have encountered a buggy <atomic> implementation.
// Also, some <atomic> implementations are much slower than a memory-barrier
// based implementation (~16ns for <atomic> based acquire-load vs. ~1ns for
// a barrier based acquire-load).
// This code is based on atomicops-internals-* in Google's perftools:
// http://code.google.com/p/google-perftools/source/browse/#svn%2Ftrunk%2Fsrc%2Fbase
#ifndef PORT_ATOMIC_POINTER_H_
#define PORT_ATOMIC_POINTER_H_
#include <stdint.h>
#ifdef LEVELDB_ATOMIC_PRESENT
#include <atomic>
#endif
#ifdef OS_WIN
#include <windows.h>
#endif
#ifdef OS_MACOSX
#include <libkern/OSAtomic.h>
#endif
#if defined(_M_X64) || defined(__x86_64__)
#define ARCH_CPU_X86_FAMILY 1
#elif defined(_M_IX86) || defined(__i386__) || defined(__i386)
#define ARCH_CPU_X86_FAMILY 1
#elif defined(__ARMEL__)
#define ARCH_CPU_ARM_FAMILY 1
#elif defined(__aarch64__)
#define ARCH_CPU_ARM64_FAMILY 1
#elif defined(__ppc__) || defined(__powerpc__) || defined(__powerpc64__)
#define ARCH_CPU_PPC_FAMILY 1
#endif
namespace leveldb {
namespace port {
// Define MemoryBarrier() if available
// Windows on x86
#if defined(OS_WIN) && defined(COMPILER_MSVC) && defined(ARCH_CPU_X86_FAMILY)
// windows.h already provides a MemoryBarrier(void) macro
// http://msdn.microsoft.com/en-us/library/ms684208(v=vs.85).aspx
#define LEVELDB_HAVE_MEMORY_BARRIER
// Mac OS
#elif defined(OS_MACOSX)
inline void MemoryBarrier() {
OSMemoryBarrier();
}
#define LEVELDB_HAVE_MEMORY_BARRIER
// Gcc on x86
#elif defined(ARCH_CPU_X86_FAMILY) && defined(__GNUC__)
inline void MemoryBarrier() {
// See http://gcc.gnu.org/ml/gcc/2003-04/msg01180.html for a discussion on
// this idiom. Also see http://en.wikipedia.org/wiki/Memory_ordering.
__asm__ __volatile__("" : : : "memory");
}
#define LEVELDB_HAVE_MEMORY_BARRIER
// Sun Studio
#elif defined(ARCH_CPU_X86_FAMILY) && defined(__SUNPRO_CC)
inline void MemoryBarrier() {
// See http://gcc.gnu.org/ml/gcc/2003-04/msg01180.html for a discussion on
// this idiom. Also see http://en.wikipedia.org/wiki/Memory_ordering.
asm volatile("" : : : "memory");
}
#define LEVELDB_HAVE_MEMORY_BARRIER
// ARM Linux
#elif defined(ARCH_CPU_ARM_FAMILY) && defined(__linux__)
typedef void (*LinuxKernelMemoryBarrierFunc)(void);
// The Linux ARM kernel provides a highly optimized device-specific memory
// barrier function at a fixed memory address that is mapped in every
// user-level process.
//
// This beats using CPU-specific instructions which are, on single-core
// devices, un-necessary and very costly (e.g. ARMv7-A "dmb" takes more
// than 180ns on a Cortex-A8 like the one on a Nexus One). Benchmarking
// shows that the extra function call cost is completely negligible on
// multi-core devices.
//
inline void MemoryBarrier() {
(*(LinuxKernelMemoryBarrierFunc)0xffff0fa0)();
}
#define LEVELDB_HAVE_MEMORY_BARRIER
// ARM64
#elif defined(ARCH_CPU_ARM64_FAMILY)
inline void MemoryBarrier() {
asm volatile("dmb sy" : : : "memory");
}
#define LEVELDB_HAVE_MEMORY_BARRIER
// PPC
#elif defined(ARCH_CPU_PPC_FAMILY) && defined(__GNUC__)
inline void MemoryBarrier() {
// TODO for some powerpc expert: is there a cheaper suitable variant?
// Perhaps by having separate barriers for acquire and release ops.
asm volatile("sync" : : : "memory");
}
#define LEVELDB_HAVE_MEMORY_BARRIER
#endif
// AtomicPointer built using platform-specific MemoryBarrier()
#if defined(LEVELDB_HAVE_MEMORY_BARRIER)
class AtomicPointer {
private:
void* rep_;
public:
AtomicPointer() { }
explicit AtomicPointer(void* p) : rep_(p) {}
inline void* NoBarrier_Load() const { return rep_; }
inline void NoBarrier_Store(void* v) { rep_ = v; }
inline void* Acquire_Load() const {
void* result = rep_;
MemoryBarrier();
return result;
}
inline void Release_Store(void* v) {
MemoryBarrier();
rep_ = v;
}
};
// AtomicPointer based on <cstdatomic>
#elif defined(LEVELDB_ATOMIC_PRESENT)
class AtomicPointer {
private:
std::atomic<void*> rep_;
public:
AtomicPointer() { }
explicit AtomicPointer(void* v) : rep_(v) { }
inline void* Acquire_Load() const {
return rep_.load(std::memory_order_acquire);
}
inline void Release_Store(void* v) {
rep_.store(v, std::memory_order_release);
}
inline void* NoBarrier_Load() const {
return rep_.load(std::memory_order_relaxed);
}
inline void NoBarrier_Store(void* v) {
rep_.store(v, std::memory_order_relaxed);
}
};
// Atomic pointer based on sparc memory barriers
#elif defined(__sparcv9) && defined(__GNUC__)
class AtomicPointer {
private:
void* rep_;
public:
AtomicPointer() { }
explicit AtomicPointer(void* v) : rep_(v) { }
inline void* Acquire_Load() const {
void* val;
__asm__ __volatile__ (
"ldx [%[rep_]], %[val] \n\t"
"membar #LoadLoad|#LoadStore \n\t"
: [val] "=r" (val)
: [rep_] "r" (&rep_)
: "memory");
return val;
}
inline void Release_Store(void* v) {
__asm__ __volatile__ (
"membar #LoadStore|#StoreStore \n\t"
"stx %[v], [%[rep_]] \n\t"
:
: [rep_] "r" (&rep_), [v] "r" (v)
: "memory");
}
inline void* NoBarrier_Load() const { return rep_; }
inline void NoBarrier_Store(void* v) { rep_ = v; }
};
// Atomic pointer based on ia64 acq/rel
#elif defined(__ia64) && defined(__GNUC__)
class AtomicPointer {
private:
void* rep_;
public:
AtomicPointer() { }
explicit AtomicPointer(void* v) : rep_(v) { }
inline void* Acquire_Load() const {
void* val ;
__asm__ __volatile__ (
"ld8.acq %[val] = [%[rep_]] \n\t"
: [val] "=r" (val)
: [rep_] "r" (&rep_)
: "memory"
);
return val;
}
inline void Release_Store(void* v) {
__asm__ __volatile__ (
"st8.rel [%[rep_]] = %[v] \n\t"
:
: [rep_] "r" (&rep_), [v] "r" (v)
: "memory"
);
}
inline void* NoBarrier_Load() const { return rep_; }
inline void NoBarrier_Store(void* v) { rep_ = v; }
};
// We have neither MemoryBarrier(), nor <atomic>
#else
#error Please implement AtomicPointer for this platform.
#endif
#undef LEVELDB_HAVE_MEMORY_BARRIER
#undef ARCH_CPU_X86_FAMILY
#undef ARCH_CPU_ARM_FAMILY
#undef ARCH_CPU_ARM64_FAMILY
#undef ARCH_CPU_PPC_FAMILY
} // namespace port
} // namespace leveldb
#endif // PORT_ATOMIC_POINTER_H_

View File

@ -10,10 +10,10 @@
// Include the appropriate platform specific file below. If you are // Include the appropriate platform specific file below. If you are
// porting to a new platform, see "port_example.h" for documentation // porting to a new platform, see "port_example.h" for documentation
// of what the new port_<platform>.h file must provide. // of what the new port_<platform>.h file must provide.
#if defined(LEVELDB_PLATFORM_POSIX) || defined(LEVELDB_PLATFORM_WINDOWS) #if defined(LEVELDB_PLATFORM_POSIX)
#include "port/port_stdcxx.h" # include "port/port_posix.h"
#elif defined(LEVELDB_PLATFORM_CHROMIUM) #elif defined(LEVELDB_PLATFORM_CHROMIUM)
#include "port/port_chromium.h" # include "port/port_chromium.h"
#endif #endif
#endif // STORAGE_LEVELDB_PORT_PORT_H_ #endif // STORAGE_LEVELDB_PORT_PORT_H_

View File

@ -1,38 +0,0 @@
// Copyright 2017 The LevelDB Authors. All rights reserved.
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file. See the AUTHORS file for names of contributors.
#ifndef STORAGE_LEVELDB_PORT_PORT_CONFIG_H_
#define STORAGE_LEVELDB_PORT_PORT_CONFIG_H_
// Define to 1 if you have a definition for fdatasync() in <unistd.h>.
#if !defined(HAVE_FDATASYNC)
#cmakedefine01 HAVE_FDATASYNC
#endif // !defined(HAVE_FDATASYNC)
// Define to 1 if you have a definition for F_FULLFSYNC in <fcntl.h>.
#if !defined(HAVE_FULLFSYNC)
#cmakedefine01 HAVE_FULLFSYNC
#endif // !defined(HAVE_FULLFSYNC)
// Define to 1 if you have a definition for O_CLOEXEC in <fcntl.h>.
#if !defined(HAVE_O_CLOEXEC)
#cmakedefine01 HAVE_O_CLOEXEC
#endif // !defined(HAVE_O_CLOEXEC)
// Define to 1 if you have Google CRC32C.
#if !defined(HAVE_CRC32C)
#cmakedefine01 HAVE_CRC32C
#endif // !defined(HAVE_CRC32C)
// Define to 1 if you have Google Snappy.
#if !defined(HAVE_SNAPPY)
#cmakedefine01 HAVE_SNAPPY
#endif // !defined(HAVE_SNAPPY)
// Define to 1 if you have Zstd.
#if !defined(HAVE_Zstd)
#cmakedefine01 HAVE_ZSTD
#endif // !defined(HAVE_ZSTD)
#endif // STORAGE_LEVELDB_PORT_PORT_CONFIG_H_

View File

@ -10,34 +10,36 @@
#ifndef STORAGE_LEVELDB_PORT_PORT_EXAMPLE_H_ #ifndef STORAGE_LEVELDB_PORT_PORT_EXAMPLE_H_
#define STORAGE_LEVELDB_PORT_PORT_EXAMPLE_H_ #define STORAGE_LEVELDB_PORT_PORT_EXAMPLE_H_
#include "port/thread_annotations.h"
namespace leveldb { namespace leveldb {
namespace port { namespace port {
// TODO(jorlow): Many of these belong more in the environment class rather than // TODO(jorlow): Many of these belong more in the environment class rather than
// here. We should try moving them and see if it affects perf. // here. We should try moving them and see if it affects perf.
// The following boolean constant must be true on a little-endian machine
// and false otherwise.
static const bool kLittleEndian = true /* or some other expression */;
// ------------------ Threading ------------------- // ------------------ Threading -------------------
// A Mutex represents an exclusive lock. // A Mutex represents an exclusive lock.
class LOCKABLE Mutex { class Mutex {
public: public:
Mutex(); Mutex();
~Mutex(); ~Mutex();
// Lock the mutex. Waits until other lockers have exited. // Lock the mutex. Waits until other lockers have exited.
// Will deadlock if the mutex is already locked by this thread. // Will deadlock if the mutex is already locked by this thread.
void Lock() EXCLUSIVE_LOCK_FUNCTION(); void Lock();
// Unlock the mutex. // Unlock the mutex.
// REQUIRES: This mutex was locked by this thread. // REQUIRES: This mutex was locked by this thread.
void Unlock() UNLOCK_FUNCTION(); void Unlock();
// Optionally crash if this thread does not hold this mutex. // Optionally crash if this thread does not hold this mutex.
// The implementation must be fast, especially if NDEBUG is // The implementation must be fast, especially if NDEBUG is
// defined. The implementation is allowed to skip all checks. // defined. The implementation is allowed to skip all checks.
void AssertHeld() ASSERT_EXCLUSIVE_LOCK(); void AssertHeld();
}; };
class CondVar { class CondVar {
@ -55,64 +57,77 @@ class CondVar {
void Signal(); void Signal();
// Wake up all waiting threads. // Wake up all waiting threads.
void SignalAll(); void SignallAll();
};
// Thread-safe initialization.
// Used as follows:
// static port::OnceType init_control = LEVELDB_ONCE_INIT;
// static void Initializer() { ... do something ...; }
// ...
// port::InitOnce(&init_control, &Initializer);
typedef intptr_t OnceType;
#define LEVELDB_ONCE_INIT 0
extern void InitOnce(port::OnceType*, void (*initializer)());
// A type that holds a pointer that can be read or written atomically
// (i.e., without word-tearing.)
class AtomicPointer {
private:
intptr_t rep_;
public:
// Initialize to arbitrary value
AtomicPointer();
// Initialize to hold v
explicit AtomicPointer(void* v) : rep_(v) { }
// Read and return the stored pointer with the guarantee that no
// later memory access (read or write) by this thread can be
// reordered ahead of this read.
void* Acquire_Load() const;
// Set v as the stored pointer with the guarantee that no earlier
// memory access (read or write) by this thread can be reordered
// after this store.
void Release_Store(void* v);
// Read the stored pointer with no ordering guarantees.
void* NoBarrier_Load() const;
// Set va as the stored pointer with no ordering guarantees.
void NoBarrier_Store(void* v);
}; };
// ------------------ Compression ------------------- // ------------------ Compression -------------------
// Store the snappy compression of "input[0,input_length-1]" in *output. // Store the snappy compression of "input[0,input_length-1]" in *output.
// Returns false if snappy is not supported by this port. // Returns false if snappy is not supported by this port.
bool Snappy_Compress(const char* input, size_t input_length, extern bool Snappy_Compress(const char* input, size_t input_length,
std::string* output); std::string* output);
// If input[0,input_length-1] looks like a valid snappy compressed // If input[0,input_length-1] looks like a valid snappy compressed
// buffer, store the size of the uncompressed data in *result and // buffer, store the size of the uncompressed data in *result and
// return true. Else return false. // return true. Else return false.
bool Snappy_GetUncompressedLength(const char* input, size_t length, extern bool Snappy_GetUncompressedLength(const char* input, size_t length,
size_t* result); size_t* result);
// Attempt to snappy uncompress input[0,input_length-1] into *output. // Attempt to snappy uncompress input[0,input_length-1] into *output.
// Returns true if successful, false if the input is invalid snappy // Returns true if successful, false if the input is invalid lightweight
// compressed data. // compressed data.
// //
// REQUIRES: at least the first "n" bytes of output[] must be writable // REQUIRES: at least the first "n" bytes of output[] must be writable
// where "n" is the result of a successful call to // where "n" is the result of a successful call to
// Snappy_GetUncompressedLength. // Snappy_GetUncompressedLength.
bool Snappy_Uncompress(const char* input_data, size_t input_length, extern bool Snappy_Uncompress(const char* input_data, size_t input_length,
char* output); char* output);
// Store the zstd compression of "input[0,input_length-1]" in *output.
// Returns false if zstd is not supported by this port.
bool Zstd_Compress(int level, const char* input, size_t input_length,
std::string* output);
// If input[0,input_length-1] looks like a valid zstd compressed
// buffer, store the size of the uncompressed data in *result and
// return true. Else return false.
bool Zstd_GetUncompressedLength(const char* input, size_t length,
size_t* result);
// Attempt to zstd uncompress input[0,input_length-1] into *output.
// Returns true if successful, false if the input is invalid zstd
// compressed data.
//
// REQUIRES: at least the first "n" bytes of output[] must be writable
// where "n" is the result of a successful call to
// Zstd_GetUncompressedLength.
bool Zstd_Uncompress(const char* input_data, size_t input_length, char* output);
// ------------------ Miscellaneous ------------------- // ------------------ Miscellaneous -------------------
// If heap profiling is not supported, returns false. // If heap profiling is not supported, returns false.
// Else repeatedly calls (*func)(arg, data, n) and then returns true. // Else repeatedly calls (*func)(arg, data, n) and then returns true.
// The concatenation of all "data[0,n-1]" fragments is the heap profile. // The concatenation of all "data[0,n-1]" fragments is the heap profile.
bool GetHeapProfile(void (*func)(void*, const char*, int), void* arg); extern bool GetHeapProfile(void (*func)(void*, const char*, int), void* arg);
// Extend the CRC to include the first n bytes of buf.
//
// Returns zero if the CRC cannot be extended using acceleration, else returns
// the newly extended CRC value (which may also be zero).
uint32_t AcceleratedCRC32C(uint32_t crc, const char* buf, size_t size);
} // namespace port } // namespace port
} // namespace leveldb } // namespace leveldb

54
port/port_posix.cc Normal file
View File

@ -0,0 +1,54 @@
// Copyright (c) 2011 The LevelDB Authors. All rights reserved.
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file. See the AUTHORS file for names of contributors.
#include "port/port_posix.h"
#include <cstdlib>
#include <stdio.h>
#include <string.h>
#include "util/logging.h"
namespace leveldb {
namespace port {
static void PthreadCall(const char* label, int result) {
if (result != 0) {
fprintf(stderr, "pthread %s: %s\n", label, strerror(result));
abort();
}
}
Mutex::Mutex() { PthreadCall("init mutex", pthread_mutex_init(&mu_, NULL)); }
Mutex::~Mutex() { PthreadCall("destroy mutex", pthread_mutex_destroy(&mu_)); }
void Mutex::Lock() { PthreadCall("lock", pthread_mutex_lock(&mu_)); }
void Mutex::Unlock() { PthreadCall("unlock", pthread_mutex_unlock(&mu_)); }
CondVar::CondVar(Mutex* mu)
: mu_(mu) {
PthreadCall("init cv", pthread_cond_init(&cv_, NULL));
}
CondVar::~CondVar() { PthreadCall("destroy cv", pthread_cond_destroy(&cv_)); }
void CondVar::Wait() {
PthreadCall("wait", pthread_cond_wait(&cv_, &mu_->mu_));
}
void CondVar::Signal() {
PthreadCall("signal", pthread_cond_signal(&cv_));
}
void CondVar::SignalAll() {
PthreadCall("broadcast", pthread_cond_broadcast(&cv_));
}
void InitOnce(OnceType* once, void (*initializer)()) {
PthreadCall("once", pthread_once(once, initializer));
}
} // namespace port
} // namespace leveldb

154
port/port_posix.h Normal file
View File

@ -0,0 +1,154 @@
// Copyright (c) 2011 The LevelDB Authors. All rights reserved.
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file. See the AUTHORS file for names of contributors.
//
// See port_example.h for documentation for the following types/functions.
#ifndef STORAGE_LEVELDB_PORT_PORT_POSIX_H_
#define STORAGE_LEVELDB_PORT_PORT_POSIX_H_
#undef PLATFORM_IS_LITTLE_ENDIAN
#if defined(OS_MACOSX)
#include <machine/endian.h>
#if defined(__DARWIN_LITTLE_ENDIAN) && defined(__DARWIN_BYTE_ORDER)
#define PLATFORM_IS_LITTLE_ENDIAN \
(__DARWIN_BYTE_ORDER == __DARWIN_LITTLE_ENDIAN)
#endif
#elif defined(OS_SOLARIS)
#include <sys/isa_defs.h>
#ifdef _LITTLE_ENDIAN
#define PLATFORM_IS_LITTLE_ENDIAN true
#else
#define PLATFORM_IS_LITTLE_ENDIAN false
#endif
#elif defined(OS_FREEBSD) || defined(OS_OPENBSD) ||\
defined(OS_NETBSD) || defined(OS_DRAGONFLYBSD)
#include <sys/types.h>
#include <sys/endian.h>
#define PLATFORM_IS_LITTLE_ENDIAN (_BYTE_ORDER == _LITTLE_ENDIAN)
#elif defined(OS_HPUX)
#define PLATFORM_IS_LITTLE_ENDIAN false
#elif defined(OS_ANDROID)
// Due to a bug in the NDK x86 <sys/endian.h> definition,
// _BYTE_ORDER must be used instead of __BYTE_ORDER on Android.
// See http://code.google.com/p/android/issues/detail?id=39824
#include <endian.h>
#define PLATFORM_IS_LITTLE_ENDIAN (_BYTE_ORDER == _LITTLE_ENDIAN)
#else
#include <endian.h>
#endif
#include <pthread.h>
#ifdef SNAPPY
#include <snappy.h>
#endif
#include <stdint.h>
#include <string>
#include "port/atomic_pointer.h"
#ifndef PLATFORM_IS_LITTLE_ENDIAN
#define PLATFORM_IS_LITTLE_ENDIAN (__BYTE_ORDER == __LITTLE_ENDIAN)
#endif
#if defined(OS_MACOSX) || defined(OS_SOLARIS) || defined(OS_FREEBSD) ||\
defined(OS_NETBSD) || defined(OS_OPENBSD) || defined(OS_DRAGONFLYBSD) ||\
defined(OS_ANDROID) || defined(OS_HPUX) || defined(CYGWIN)
// Use fread/fwrite/fflush on platforms without _unlocked variants
#define fread_unlocked fread
#define fwrite_unlocked fwrite
#define fflush_unlocked fflush
#endif
#if defined(OS_MACOSX) || defined(OS_FREEBSD) ||\
defined(OS_OPENBSD) || defined(OS_DRAGONFLYBSD)
// Use fsync() on platforms without fdatasync()
#define fdatasync fsync
#endif
#if defined(OS_ANDROID) && __ANDROID_API__ < 9
// fdatasync() was only introduced in API level 9 on Android. Use fsync()
// when targetting older platforms.
#define fdatasync fsync
#endif
namespace leveldb {
namespace port {
static const bool kLittleEndian = PLATFORM_IS_LITTLE_ENDIAN;
#undef PLATFORM_IS_LITTLE_ENDIAN
class CondVar;
class Mutex {
public:
Mutex();
~Mutex();
void Lock();
void Unlock();
void AssertHeld() { }
private:
friend class CondVar;
pthread_mutex_t mu_;
// No copying
Mutex(const Mutex&);
void operator=(const Mutex&);
};
class CondVar {
public:
explicit CondVar(Mutex* mu);
~CondVar();
void Wait();
void Signal();
void SignalAll();
private:
pthread_cond_t cv_;
Mutex* mu_;
};
typedef pthread_once_t OnceType;
#define LEVELDB_ONCE_INIT PTHREAD_ONCE_INIT
extern void InitOnce(OnceType* once, void (*initializer)());
inline bool Snappy_Compress(const char* input, size_t length,
::std::string* output) {
#ifdef SNAPPY
output->resize(snappy::MaxCompressedLength(length));
size_t outlen;
snappy::RawCompress(input, length, &(*output)[0], &outlen);
output->resize(outlen);
return true;
#endif
return false;
}
inline bool Snappy_GetUncompressedLength(const char* input, size_t length,
size_t* result) {
#ifdef SNAPPY
return snappy::GetUncompressedLength(input, length, result);
#else
return false;
#endif
}
inline bool Snappy_Uncompress(const char* input, size_t length,
char* output) {
#ifdef SNAPPY
return snappy::RawUncompress(input, length, output);
#else
return false;
#endif
}
inline bool GetHeapProfile(void (*func)(void*, const char*, int), void* arg) {
return false;
}
} // namespace port
} // namespace leveldb
#endif // STORAGE_LEVELDB_PORT_PORT_POSIX_H_

View File

@ -1,223 +0,0 @@
// Copyright (c) 2018 The LevelDB Authors. All rights reserved.
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file. See the AUTHORS file for names of contributors.
#ifndef STORAGE_LEVELDB_PORT_PORT_STDCXX_H_
#define STORAGE_LEVELDB_PORT_PORT_STDCXX_H_
// port/port_config.h availability is automatically detected via __has_include
// in newer compilers. If LEVELDB_HAS_PORT_CONFIG_H is defined, it overrides the
// configuration detection.
#if defined(LEVELDB_HAS_PORT_CONFIG_H)
#if LEVELDB_HAS_PORT_CONFIG_H
#include "port/port_config.h"
#endif // LEVELDB_HAS_PORT_CONFIG_H
#elif defined(__has_include)
#if __has_include("port/port_config.h")
#include "port/port_config.h"
#endif // __has_include("port/port_config.h")
#endif // defined(LEVELDB_HAS_PORT_CONFIG_H)
#if HAVE_CRC32C
#include <crc32c/crc32c.h>
#endif // HAVE_CRC32C
#if HAVE_SNAPPY
#include <snappy.h>
#endif // HAVE_SNAPPY
#if HAVE_ZSTD
#define ZSTD_STATIC_LINKING_ONLY // For ZSTD_compressionParameters.
#include <zstd.h>
#endif // HAVE_ZSTD
#include <cassert>
#include <condition_variable> // NOLINT
#include <cstddef>
#include <cstdint>
#include <mutex> // NOLINT
#include <string>
#include "port/thread_annotations.h"
namespace leveldb {
namespace port {
class CondVar;
// Thinly wraps std::mutex.
class LOCKABLE Mutex {
public:
Mutex() = default;
~Mutex() = default;
Mutex(const Mutex&) = delete;
Mutex& operator=(const Mutex&) = delete;
void Lock() EXCLUSIVE_LOCK_FUNCTION() { mu_.lock(); }
void Unlock() UNLOCK_FUNCTION() { mu_.unlock(); }
void AssertHeld() ASSERT_EXCLUSIVE_LOCK() {}
private:
friend class CondVar;
std::mutex mu_;
};
// Thinly wraps std::condition_variable.
class CondVar {
public:
explicit CondVar(Mutex* mu) : mu_(mu) { assert(mu != nullptr); }
~CondVar() = default;
CondVar(const CondVar&) = delete;
CondVar& operator=(const CondVar&) = delete;
void Wait() {
std::unique_lock<std::mutex> lock(mu_->mu_, std::adopt_lock);
cv_.wait(lock);
lock.release();
}
void Signal() { cv_.notify_one(); }
void SignalAll() { cv_.notify_all(); }
private:
std::condition_variable cv_;
Mutex* const mu_;
};
inline bool Snappy_Compress(const char* input, size_t length,
std::string* output) {
#if HAVE_SNAPPY
output->resize(snappy::MaxCompressedLength(length));
size_t outlen;
snappy::RawCompress(input, length, &(*output)[0], &outlen);
output->resize(outlen);
return true;
#else
// Silence compiler warnings about unused arguments.
(void)input;
(void)length;
(void)output;
#endif // HAVE_SNAPPY
return false;
}
inline bool Snappy_GetUncompressedLength(const char* input, size_t length,
size_t* result) {
#if HAVE_SNAPPY
return snappy::GetUncompressedLength(input, length, result);
#else
// Silence compiler warnings about unused arguments.
(void)input;
(void)length;
(void)result;
return false;
#endif // HAVE_SNAPPY
}
inline bool Snappy_Uncompress(const char* input, size_t length, char* output) {
#if HAVE_SNAPPY
return snappy::RawUncompress(input, length, output);
#else
// Silence compiler warnings about unused arguments.
(void)input;
(void)length;
(void)output;
return false;
#endif // HAVE_SNAPPY
}
inline bool Zstd_Compress(int level, const char* input, size_t length,
std::string* output) {
#if HAVE_ZSTD
// Get the MaxCompressedLength.
size_t outlen = ZSTD_compressBound(length);
if (ZSTD_isError(outlen)) {
return false;
}
output->resize(outlen);
ZSTD_CCtx* ctx = ZSTD_createCCtx();
ZSTD_compressionParameters parameters =
ZSTD_getCParams(level, std::max(length, size_t{1}), /*dictSize=*/0);
ZSTD_CCtx_setCParams(ctx, parameters);
outlen = ZSTD_compress2(ctx, &(*output)[0], output->size(), input, length);
ZSTD_freeCCtx(ctx);
if (ZSTD_isError(outlen)) {
return false;
}
output->resize(outlen);
return true;
#else
// Silence compiler warnings about unused arguments.
(void)level;
(void)input;
(void)length;
(void)output;
return false;
#endif // HAVE_ZSTD
}
inline bool Zstd_GetUncompressedLength(const char* input, size_t length,
size_t* result) {
#if HAVE_ZSTD
size_t size = ZSTD_getFrameContentSize(input, length);
if (size == 0) return false;
*result = size;
return true;
#else
// Silence compiler warnings about unused arguments.
(void)input;
(void)length;
(void)result;
return false;
#endif // HAVE_ZSTD
}
inline bool Zstd_Uncompress(const char* input, size_t length, char* output) {
#if HAVE_ZSTD
size_t outlen;
if (!Zstd_GetUncompressedLength(input, length, &outlen)) {
return false;
}
ZSTD_DCtx* ctx = ZSTD_createDCtx();
outlen = ZSTD_decompressDCtx(ctx, output, outlen, input, length);
ZSTD_freeDCtx(ctx);
if (ZSTD_isError(outlen)) {
return false;
}
return true;
#else
// Silence compiler warnings about unused arguments.
(void)input;
(void)length;
(void)output;
return false;
#endif // HAVE_ZSTD
}
inline bool GetHeapProfile(void (*func)(void*, const char*, int), void* arg) {
// Silence compiler warnings about unused arguments.
(void)func;
(void)arg;
return false;
}
inline uint32_t AcceleratedCRC32C(uint32_t crc, const char* buf, size_t size) {
#if HAVE_CRC32C
return ::crc32c::Extend(crc, reinterpret_cast<const uint8_t*>(buf), size);
#else
// Silence compiler warnings about unused arguments.
(void)crc;
(void)buf;
(void)size;
return 0;
#endif // HAVE_CRC32C
}
} // namespace port
} // namespace leveldb
#endif // STORAGE_LEVELDB_PORT_PORT_STDCXX_H_

View File

@ -5,104 +5,56 @@
#ifndef STORAGE_LEVELDB_PORT_THREAD_ANNOTATIONS_H_ #ifndef STORAGE_LEVELDB_PORT_THREAD_ANNOTATIONS_H_
#define STORAGE_LEVELDB_PORT_THREAD_ANNOTATIONS_H_ #define STORAGE_LEVELDB_PORT_THREAD_ANNOTATIONS_H_
// Use Clang's thread safety analysis annotations when available. In other // Some environments provide custom macros to aid in static thread-safety
// environments, the macros receive empty definitions. // analysis. Provide empty definitions of such macros unless they are already
// Usage documentation: https://clang.llvm.org/docs/ThreadSafetyAnalysis.html // defined.
#if !defined(THREAD_ANNOTATION_ATTRIBUTE__)
#if defined(__clang__)
#define THREAD_ANNOTATION_ATTRIBUTE__(x) __attribute__((x))
#else
#define THREAD_ANNOTATION_ATTRIBUTE__(x) // no-op
#endif
#endif // !defined(THREAD_ANNOTATION_ATTRIBUTE__)
#ifndef GUARDED_BY
#define GUARDED_BY(x) THREAD_ANNOTATION_ATTRIBUTE__(guarded_by(x))
#endif
#ifndef PT_GUARDED_BY
#define PT_GUARDED_BY(x) THREAD_ANNOTATION_ATTRIBUTE__(pt_guarded_by(x))
#endif
#ifndef ACQUIRED_AFTER
#define ACQUIRED_AFTER(...) \
THREAD_ANNOTATION_ATTRIBUTE__(acquired_after(__VA_ARGS__))
#endif
#ifndef ACQUIRED_BEFORE
#define ACQUIRED_BEFORE(...) \
THREAD_ANNOTATION_ATTRIBUTE__(acquired_before(__VA_ARGS__))
#endif
#ifndef EXCLUSIVE_LOCKS_REQUIRED #ifndef EXCLUSIVE_LOCKS_REQUIRED
#define EXCLUSIVE_LOCKS_REQUIRED(...) \ #define EXCLUSIVE_LOCKS_REQUIRED(...)
THREAD_ANNOTATION_ATTRIBUTE__(exclusive_locks_required(__VA_ARGS__))
#endif #endif
#ifndef SHARED_LOCKS_REQUIRED #ifndef SHARED_LOCKS_REQUIRED
#define SHARED_LOCKS_REQUIRED(...) \ #define SHARED_LOCKS_REQUIRED(...)
THREAD_ANNOTATION_ATTRIBUTE__(shared_locks_required(__VA_ARGS__))
#endif #endif
#ifndef LOCKS_EXCLUDED #ifndef LOCKS_EXCLUDED
#define LOCKS_EXCLUDED(...) \ #define LOCKS_EXCLUDED(...)
THREAD_ANNOTATION_ATTRIBUTE__(locks_excluded(__VA_ARGS__))
#endif #endif
#ifndef LOCK_RETURNED #ifndef LOCK_RETURNED
#define LOCK_RETURNED(x) THREAD_ANNOTATION_ATTRIBUTE__(lock_returned(x)) #define LOCK_RETURNED(x)
#endif #endif
#ifndef LOCKABLE #ifndef LOCKABLE
#define LOCKABLE THREAD_ANNOTATION_ATTRIBUTE__(lockable) #define LOCKABLE
#endif #endif
#ifndef SCOPED_LOCKABLE #ifndef SCOPED_LOCKABLE
#define SCOPED_LOCKABLE THREAD_ANNOTATION_ATTRIBUTE__(scoped_lockable) #define SCOPED_LOCKABLE
#endif #endif
#ifndef EXCLUSIVE_LOCK_FUNCTION #ifndef EXCLUSIVE_LOCK_FUNCTION
#define EXCLUSIVE_LOCK_FUNCTION(...) \ #define EXCLUSIVE_LOCK_FUNCTION(...)
THREAD_ANNOTATION_ATTRIBUTE__(exclusive_lock_function(__VA_ARGS__))
#endif #endif
#ifndef SHARED_LOCK_FUNCTION #ifndef SHARED_LOCK_FUNCTION
#define SHARED_LOCK_FUNCTION(...) \ #define SHARED_LOCK_FUNCTION(...)
THREAD_ANNOTATION_ATTRIBUTE__(shared_lock_function(__VA_ARGS__))
#endif #endif
#ifndef EXCLUSIVE_TRYLOCK_FUNCTION #ifndef EXCLUSIVE_TRYLOCK_FUNCTION
#define EXCLUSIVE_TRYLOCK_FUNCTION(...) \ #define EXCLUSIVE_TRYLOCK_FUNCTION(...)
THREAD_ANNOTATION_ATTRIBUTE__(exclusive_trylock_function(__VA_ARGS__))
#endif #endif
#ifndef SHARED_TRYLOCK_FUNCTION #ifndef SHARED_TRYLOCK_FUNCTION
#define SHARED_TRYLOCK_FUNCTION(...) \ #define SHARED_TRYLOCK_FUNCTION(...)
THREAD_ANNOTATION_ATTRIBUTE__(shared_trylock_function(__VA_ARGS__))
#endif #endif
#ifndef UNLOCK_FUNCTION #ifndef UNLOCK_FUNCTION
#define UNLOCK_FUNCTION(...) \ #define UNLOCK_FUNCTION(...)
THREAD_ANNOTATION_ATTRIBUTE__(unlock_function(__VA_ARGS__))
#endif #endif
#ifndef NO_THREAD_SAFETY_ANALYSIS #ifndef NO_THREAD_SAFETY_ANALYSIS
#define NO_THREAD_SAFETY_ANALYSIS \ #define NO_THREAD_SAFETY_ANALYSIS
THREAD_ANNOTATION_ATTRIBUTE__(no_thread_safety_analysis)
#endif
#ifndef ASSERT_EXCLUSIVE_LOCK
#define ASSERT_EXCLUSIVE_LOCK(...) \
THREAD_ANNOTATION_ATTRIBUTE__(assert_exclusive_lock(__VA_ARGS__))
#endif
#ifndef ASSERT_SHARED_LOCK
#define ASSERT_SHARED_LOCK(...) \
THREAD_ANNOTATION_ATTRIBUTE__(assert_shared_lock(__VA_ARGS__))
#endif #endif
#endif // STORAGE_LEVELDB_PORT_THREAD_ANNOTATIONS_H_ #endif // STORAGE_LEVELDB_PORT_THREAD_ANNOTATIONS_H_

24
port/win/stdint.h Normal file
View File

@ -0,0 +1,24 @@
// Copyright (c) 2011 The LevelDB Authors. All rights reserved.
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file. See the AUTHORS file for names of contributors.
// MSVC didn't ship with this file until the 2010 version.
#ifndef STORAGE_LEVELDB_PORT_WIN_STDINT_H_
#define STORAGE_LEVELDB_PORT_WIN_STDINT_H_
#if !defined(_MSC_VER)
#error This file should only be included when compiling with MSVC.
#endif
// Define C99 equivalent types.
typedef signed char int8_t;
typedef signed short int16_t;
typedef signed int int32_t;
typedef signed long long int64_t;
typedef unsigned char uint8_t;
typedef unsigned short uint16_t;
typedef unsigned int uint32_t;
typedef unsigned long long uint64_t;
#endif // STORAGE_LEVELDB_PORT_WIN_STDINT_H_

View File

@ -6,10 +6,8 @@
#include "table/block.h" #include "table/block.h"
#include <algorithm>
#include <cstdint>
#include <vector> #include <vector>
#include <algorithm>
#include "leveldb/comparator.h" #include "leveldb/comparator.h"
#include "table/format.h" #include "table/format.h"
#include "util/coding.h" #include "util/coding.h"
@ -29,7 +27,7 @@ Block::Block(const BlockContents& contents)
if (size_ < sizeof(uint32_t)) { if (size_ < sizeof(uint32_t)) {
size_ = 0; // Error marker size_ = 0; // Error marker
} else { } else {
size_t max_restarts_allowed = (size_ - sizeof(uint32_t)) / sizeof(uint32_t); size_t max_restarts_allowed = (size_-sizeof(uint32_t)) / sizeof(uint32_t);
if (NumRestarts() > max_restarts_allowed) { if (NumRestarts() > max_restarts_allowed) {
// The size is too small for NumRestarts() // The size is too small for NumRestarts()
size_ = 0; size_ = 0;
@ -50,26 +48,27 @@ Block::~Block() {
// and the length of the value in "*shared", "*non_shared", and // and the length of the value in "*shared", "*non_shared", and
// "*value_length", respectively. Will not dereference past "limit". // "*value_length", respectively. Will not dereference past "limit".
// //
// If any errors are detected, returns nullptr. Otherwise, returns a // If any errors are detected, returns NULL. Otherwise, returns a
// pointer to the key delta (just past the three decoded values). // pointer to the key delta (just past the three decoded values).
static inline const char* DecodeEntry(const char* p, const char* limit, static inline const char* DecodeEntry(const char* p, const char* limit,
uint32_t* shared, uint32_t* non_shared, uint32_t* shared,
uint32_t* non_shared,
uint32_t* value_length) { uint32_t* value_length) {
if (limit - p < 3) return nullptr; if (limit - p < 3) return NULL;
*shared = reinterpret_cast<const uint8_t*>(p)[0]; *shared = reinterpret_cast<const unsigned char*>(p)[0];
*non_shared = reinterpret_cast<const uint8_t*>(p)[1]; *non_shared = reinterpret_cast<const unsigned char*>(p)[1];
*value_length = reinterpret_cast<const uint8_t*>(p)[2]; *value_length = reinterpret_cast<const unsigned char*>(p)[2];
if ((*shared | *non_shared | *value_length) < 128) { if ((*shared | *non_shared | *value_length) < 128) {
// Fast path: all three values are encoded in one byte each // Fast path: all three values are encoded in one byte each
p += 3; p += 3;
} else { } else {
if ((p = GetVarint32Ptr(p, limit, shared)) == nullptr) return nullptr; if ((p = GetVarint32Ptr(p, limit, shared)) == NULL) return NULL;
if ((p = GetVarint32Ptr(p, limit, non_shared)) == nullptr) return nullptr; if ((p = GetVarint32Ptr(p, limit, non_shared)) == NULL) return NULL;
if ((p = GetVarint32Ptr(p, limit, value_length)) == nullptr) return nullptr; if ((p = GetVarint32Ptr(p, limit, value_length)) == NULL) return NULL;
} }
if (static_cast<uint32_t>(limit - p) < (*non_shared + *value_length)) { if (static_cast<uint32_t>(limit - p) < (*non_shared + *value_length)) {
return nullptr; return NULL;
} }
return p; return p;
} }
@ -77,9 +76,9 @@ static inline const char* DecodeEntry(const char* p, const char* limit,
class Block::Iter : public Iterator { class Block::Iter : public Iterator {
private: private:
const Comparator* const comparator_; const Comparator* const comparator_;
const char* const data_; // underlying block contents const char* const data_; // underlying block contents
uint32_t const restarts_; // Offset of restart array (list of fixed32) uint32_t const restarts_; // Offset of restart array (list of fixed32)
uint32_t const num_restarts_; // Number of uint32_t entries in restart array uint32_t const num_restarts_; // Number of uint32_t entries in restart array
// current_ is offset in data_ of current entry. >= restarts_ if !Valid // current_ is offset in data_ of current entry. >= restarts_ if !Valid
uint32_t current_; uint32_t current_;
@ -113,7 +112,9 @@ class Block::Iter : public Iterator {
} }
public: public:
Iter(const Comparator* comparator, const char* data, uint32_t restarts, Iter(const Comparator* comparator,
const char* data,
uint32_t restarts,
uint32_t num_restarts) uint32_t num_restarts)
: comparator_(comparator), : comparator_(comparator),
data_(data), data_(data),
@ -124,23 +125,23 @@ class Block::Iter : public Iterator {
assert(num_restarts_ > 0); assert(num_restarts_ > 0);
} }
bool Valid() const override { return current_ < restarts_; } virtual bool Valid() const { return current_ < restarts_; }
Status status() const override { return status_; } virtual Status status() const { return status_; }
Slice key() const override { virtual Slice key() const {
assert(Valid()); assert(Valid());
return key_; return key_;
} }
Slice value() const override { virtual Slice value() const {
assert(Valid()); assert(Valid());
return value_; return value_;
} }
void Next() override { virtual void Next() {
assert(Valid()); assert(Valid());
ParseNextKey(); ParseNextKey();
} }
void Prev() override { virtual void Prev() {
assert(Valid()); assert(Valid());
// Scan backwards to a restart point before current_ // Scan backwards to a restart point before current_
@ -161,37 +162,19 @@ class Block::Iter : public Iterator {
} while (ParseNextKey() && NextEntryOffset() < original); } while (ParseNextKey() && NextEntryOffset() < original);
} }
void Seek(const Slice& target) override { virtual void Seek(const Slice& target) {
// Binary search in restart array to find the last restart point // Binary search in restart array to find the last restart point
// with a key < target // with a key < target
uint32_t left = 0; uint32_t left = 0;
uint32_t right = num_restarts_ - 1; uint32_t right = num_restarts_ - 1;
int current_key_compare = 0;
if (Valid()) {
// If we're already scanning, use the current position as a starting
// point. This is beneficial if the key we're seeking to is ahead of the
// current position.
current_key_compare = Compare(key_, target);
if (current_key_compare < 0) {
// key_ is smaller than target
left = restart_index_;
} else if (current_key_compare > 0) {
right = restart_index_;
} else {
// We're seeking to the key we're already at.
return;
}
}
while (left < right) { while (left < right) {
uint32_t mid = (left + right + 1) / 2; uint32_t mid = (left + right + 1) / 2;
uint32_t region_offset = GetRestartPoint(mid); uint32_t region_offset = GetRestartPoint(mid);
uint32_t shared, non_shared, value_length; uint32_t shared, non_shared, value_length;
const char* key_ptr = const char* key_ptr = DecodeEntry(data_ + region_offset,
DecodeEntry(data_ + region_offset, data_ + restarts_, &shared, data_ + restarts_,
&non_shared, &value_length); &shared, &non_shared, &value_length);
if (key_ptr == nullptr || (shared != 0)) { if (key_ptr == NULL || (shared != 0)) {
CorruptionError(); CorruptionError();
return; return;
} }
@ -207,15 +190,8 @@ class Block::Iter : public Iterator {
} }
} }
// We might be able to use our current position within the restart block.
// This is true if we determined the key we desire is in the current block
// and is after than the current key.
assert(current_key_compare == 0 || Valid());
bool skip_seek = left == restart_index_ && current_key_compare < 0;
if (!skip_seek) {
SeekToRestartPoint(left);
}
// Linear search (within restart block) for first key >= target // Linear search (within restart block) for first key >= target
SeekToRestartPoint(left);
while (true) { while (true) {
if (!ParseNextKey()) { if (!ParseNextKey()) {
return; return;
@ -226,12 +202,12 @@ class Block::Iter : public Iterator {
} }
} }
void SeekToFirst() override { virtual void SeekToFirst() {
SeekToRestartPoint(0); SeekToRestartPoint(0);
ParseNextKey(); ParseNextKey();
} }
void SeekToLast() override { virtual void SeekToLast() {
SeekToRestartPoint(num_restarts_ - 1); SeekToRestartPoint(num_restarts_ - 1);
while (ParseNextKey() && NextEntryOffset() < restarts_) { while (ParseNextKey() && NextEntryOffset() < restarts_) {
// Keep skipping // Keep skipping
@ -261,7 +237,7 @@ class Block::Iter : public Iterator {
// Decode next entry // Decode next entry
uint32_t shared, non_shared, value_length; uint32_t shared, non_shared, value_length;
p = DecodeEntry(p, limit, &shared, &non_shared, &value_length); p = DecodeEntry(p, limit, &shared, &non_shared, &value_length);
if (p == nullptr || key_.size() < shared) { if (p == NULL || key_.size() < shared) {
CorruptionError(); CorruptionError();
return false; return false;
} else { } else {
@ -277,7 +253,7 @@ class Block::Iter : public Iterator {
} }
}; };
Iterator* Block::NewIterator(const Comparator* comparator) { Iterator* Block::NewIterator(const Comparator* cmp) {
if (size_ < sizeof(uint32_t)) { if (size_ < sizeof(uint32_t)) {
return NewErrorIterator(Status::Corruption("bad block contents")); return NewErrorIterator(Status::Corruption("bad block contents"));
} }
@ -285,7 +261,7 @@ Iterator* Block::NewIterator(const Comparator* comparator) {
if (num_restarts == 0) { if (num_restarts == 0) {
return NewEmptyIterator(); return NewEmptyIterator();
} else { } else {
return new Iter(comparator, data_, restart_offset_, num_restarts); return new Iter(cmp, data_, restart_offset_, num_restarts);
} }
} }

Some files were not shown because too many files have changed in this diff Show More