mirror of
https://github.com/eledio-devices/thirdparty-littlefs.git
synced 2025-11-01 16:14:13 +01:00
Compare commits
30 Commits
fix-lfs-mi
...
fix-migrat
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
3f99e73841 | ||
|
|
6b65737715 | ||
|
|
4ebe6030c5 | ||
|
|
7ae8d778f1 | ||
|
|
4d068a154d | ||
|
|
ba088aa213 | ||
|
|
955b296bcc | ||
|
|
241dbc6f86 | ||
|
|
8cca58f1a6 | ||
|
|
97f86af4e9 | ||
|
|
d40302c5e3 | ||
|
|
0b5a78e2cd | ||
|
|
27b6cc829b | ||
|
|
fd204ac2fb | ||
|
|
bd99402d9a | ||
|
|
bce442a86b | ||
|
|
f26e970a0e | ||
|
|
965d29b887 | ||
|
|
f7fd7d966a | ||
|
|
d5aba27d60 | ||
|
|
0c77123eee | ||
|
|
494dd6673d | ||
|
|
fce2569005 | ||
|
|
9d1f1211a9 | ||
|
|
151104c790 | ||
|
|
303ffb2da4 | ||
|
|
5bf71fa43e | ||
|
|
55fb1416c7 | ||
|
|
dc031ce1d9 | ||
|
|
f85ff1d2f8 |
26
.travis.yml
26
.travis.yml
@@ -248,34 +248,30 @@ jobs:
|
|||||||
-m "Generated v$LFS_VERSION_MAJOR prefixes")
|
-m "Generated v$LFS_VERSION_MAJOR prefixes")
|
||||||
git reset --hard
|
git reset --hard
|
||||||
# Update major version branches (vN and vN-prefix)
|
# Update major version branches (vN and vN-prefix)
|
||||||
git push https://$GEKY_BOT_RELEASES@github.com/$TRAVIS_REPO_SLUG.git \
|
git push --atomic https://$GEKY_BOT_RELEASES@github.com/$TRAVIS_REPO_SLUG.git \
|
||||||
v$LFS_VERSION_MAJOR \
|
v$LFS_VERSION_MAJOR \
|
||||||
v$LFS_VERSION_MAJOR-prefix
|
v$LFS_VERSION_MAJOR-prefix
|
||||||
# Create patch version tag (vN.N.N)
|
|
||||||
curl -f -u "$GEKY_BOT_RELEASES" -X POST \
|
|
||||||
https://api.github.com/repos/$TRAVIS_REPO_SLUG/git/refs \
|
|
||||||
-d "{
|
|
||||||
\"ref\": \"refs/tags/$LFS_VERSION\",
|
|
||||||
\"sha\": \"$TRAVIS_COMMIT\"
|
|
||||||
}"
|
|
||||||
# Create minor release?
|
|
||||||
[[ "$LFS_VERSION" == *.0 ]] || exit 0
|
|
||||||
# Build release notes
|
# Build release notes
|
||||||
PREV=$(git tag --sort=-v:refname -l "v*.0" | head -1)
|
PREV=$(git tag --sort=-v:refname -l "v*" | head -1)
|
||||||
if [ ! -z "$PREV" ]
|
if [ ! -z "$PREV" ]
|
||||||
then
|
then
|
||||||
echo "PREV $PREV"
|
echo "PREV $PREV"
|
||||||
CHANGES=$'### Changes\n\n'$( \
|
CHANGES=$(git log --oneline $PREV.. --grep='^Merge' --invert-grep)
|
||||||
git log --oneline $PREV.. --grep='^Merge' --invert-grep)
|
|
||||||
printf "CHANGES\n%s\n\n" "$CHANGES"
|
printf "CHANGES\n%s\n\n" "$CHANGES"
|
||||||
fi
|
fi
|
||||||
# Create the release
|
case ${GEKY_BOT_DRAFT:-minor} in
|
||||||
|
true) DRAFT=true ;;
|
||||||
|
minor) DRAFT=$(jq -R 'endswith(".0")' <<< "$LFS_VERSION") ;;
|
||||||
|
false) DRAFT=false ;;
|
||||||
|
esac
|
||||||
|
# Create the release and patch version tag (vN.N.N)
|
||||||
curl -f -u "$GEKY_BOT_RELEASES" -X POST \
|
curl -f -u "$GEKY_BOT_RELEASES" -X POST \
|
||||||
https://api.github.com/repos/$TRAVIS_REPO_SLUG/releases \
|
https://api.github.com/repos/$TRAVIS_REPO_SLUG/releases \
|
||||||
-d "{
|
-d "{
|
||||||
\"tag_name\": \"$LFS_VERSION\",
|
\"tag_name\": \"$LFS_VERSION\",
|
||||||
\"name\": \"${LFS_VERSION%.0}\",
|
\"name\": \"${LFS_VERSION%.0}\",
|
||||||
\"draft\": true,
|
\"target_commitish\": \"$TRAVIS_COMMIT\",
|
||||||
|
\"draft\": $DRAFT,
|
||||||
\"body\": $(jq -sR '.' <<< "$CHANGES")
|
\"body\": $(jq -sR '.' <<< "$CHANGES")
|
||||||
}" #"
|
}" #"
|
||||||
SCRIPT
|
SCRIPT
|
||||||
|
|||||||
54
DESIGN.md
54
DESIGN.md
@@ -254,7 +254,7 @@ have weaknesses that limit their usefulness. But if we merge the two they can
|
|||||||
mutually solve each other's limitations.
|
mutually solve each other's limitations.
|
||||||
|
|
||||||
This is the idea behind littlefs. At the sub-block level, littlefs is built
|
This is the idea behind littlefs. At the sub-block level, littlefs is built
|
||||||
out of small, two blocks logs that provide atomic updates to metadata anywhere
|
out of small, two block logs that provide atomic updates to metadata anywhere
|
||||||
on the filesystem. At the super-block level, littlefs is a CObW tree of blocks
|
on the filesystem. At the super-block level, littlefs is a CObW tree of blocks
|
||||||
that can be evicted on demand.
|
that can be evicted on demand.
|
||||||
|
|
||||||
@@ -676,7 +676,7 @@ block, this cost is fairly reasonable.
|
|||||||
---
|
---
|
||||||
|
|
||||||
This is a new data structure, so we still have several questions. What is the
|
This is a new data structure, so we still have several questions. What is the
|
||||||
storage overage? Can the number of pointers exceed the size of a block? How do
|
storage overhead? Can the number of pointers exceed the size of a block? How do
|
||||||
we store a CTZ skip-list in our metadata pairs?
|
we store a CTZ skip-list in our metadata pairs?
|
||||||
|
|
||||||
To find the storage overhead, we can look at the data structure as multiple
|
To find the storage overhead, we can look at the data structure as multiple
|
||||||
@@ -742,8 +742,8 @@ where:
|
|||||||
2. popcount(![x]) = the number of bits that are 1 in ![x]
|
2. popcount(![x]) = the number of bits that are 1 in ![x]
|
||||||
|
|
||||||
Initial tests of this surprising property seem to hold. As ![n] approaches
|
Initial tests of this surprising property seem to hold. As ![n] approaches
|
||||||
infinity, we end up with an average overhead of 2 pointers, which matches what
|
infinity, we end up with an average overhead of 2 pointers, which matches our
|
||||||
our assumption from earlier. During iteration, the popcount function seems to
|
assumption from earlier. During iteration, the popcount function seems to
|
||||||
handle deviations from this average. Of course, just to make sure I wrote a
|
handle deviations from this average. Of course, just to make sure I wrote a
|
||||||
quick script that verified this property for all 32-bit integers.
|
quick script that verified this property for all 32-bit integers.
|
||||||
|
|
||||||
@@ -767,7 +767,7 @@ overflow, but we can avoid this by rearranging the equation a bit:
|
|||||||
|
|
||||||
![off = N - (B-2w/8)n - (w/8)popcount(n)][ctz-formula7]
|
![off = N - (B-2w/8)n - (w/8)popcount(n)][ctz-formula7]
|
||||||
|
|
||||||
Our solution requires quite a bit of math, but computer are very good at math.
|
Our solution requires quite a bit of math, but computers are very good at math.
|
||||||
Now we can find both our block index and offset from a size in _O(1)_, letting
|
Now we can find both our block index and offset from a size in _O(1)_, letting
|
||||||
us store CTZ skip-lists with only a pointer and size.
|
us store CTZ skip-lists with only a pointer and size.
|
||||||
|
|
||||||
@@ -850,7 +850,7 @@ nearly every write to the filesystem.
|
|||||||
|
|
||||||
Normally, block allocation involves some sort of free list or bitmap stored on
|
Normally, block allocation involves some sort of free list or bitmap stored on
|
||||||
the filesystem that is updated with free blocks. However, with power
|
the filesystem that is updated with free blocks. However, with power
|
||||||
resilience, keeping these structure consistent becomes difficult. It doesn't
|
resilience, keeping these structures consistent becomes difficult. It doesn't
|
||||||
help that any mistake in updating these structures can result in lost blocks
|
help that any mistake in updating these structures can result in lost blocks
|
||||||
that are impossible to recover.
|
that are impossible to recover.
|
||||||
|
|
||||||
@@ -894,9 +894,9 @@ high-risk error conditions.
|
|||||||
---
|
---
|
||||||
|
|
||||||
Our block allocator needs to find free blocks efficiently. You could traverse
|
Our block allocator needs to find free blocks efficiently. You could traverse
|
||||||
through every block on storage and check each one against our filesystem tree,
|
through every block on storage and check each one against our filesystem tree;
|
||||||
however the runtime would be abhorrent. We need to somehow collect multiple
|
however, the runtime would be abhorrent. We need to somehow collect multiple
|
||||||
blocks each traversal.
|
blocks per traversal.
|
||||||
|
|
||||||
Looking at existing designs, some larger filesystems that use a similar "drop
|
Looking at existing designs, some larger filesystems that use a similar "drop
|
||||||
it on the floor" strategy store a bitmap of the entire storage in [RAM]. This
|
it on the floor" strategy store a bitmap of the entire storage in [RAM]. This
|
||||||
@@ -920,8 +920,8 @@ a brute force traversal. Instead of a bitmap the size of storage, we keep track
|
|||||||
of a small, fixed-size bitmap called the lookahead buffer. During block
|
of a small, fixed-size bitmap called the lookahead buffer. During block
|
||||||
allocation, we take blocks from the lookahead buffer. If the lookahead buffer
|
allocation, we take blocks from the lookahead buffer. If the lookahead buffer
|
||||||
is empty, we scan the filesystem for more free blocks, populating our lookahead
|
is empty, we scan the filesystem for more free blocks, populating our lookahead
|
||||||
buffer. Each scan we use an increasing offset, circling the storage as blocks
|
buffer. In each scan we use an increasing offset, circling the storage as
|
||||||
are allocated.
|
blocks are allocated.
|
||||||
|
|
||||||
Here's what it might look like to allocate 4 blocks on a decently busy
|
Here's what it might look like to allocate 4 blocks on a decently busy
|
||||||
filesystem with a 32 bit lookahead and a total of 128 blocks (512 KiB
|
filesystem with a 32 bit lookahead and a total of 128 blocks (512 KiB
|
||||||
@@ -950,7 +950,7 @@ alloc = 112 lookahead: ffff8000
|
|||||||
```
|
```
|
||||||
|
|
||||||
This lookahead approach has a runtime complexity of _O(n²)_ to completely
|
This lookahead approach has a runtime complexity of _O(n²)_ to completely
|
||||||
scan storage, however, bitmaps are surprisingly compact, and in practice only
|
scan storage; however, bitmaps are surprisingly compact, and in practice only
|
||||||
one or two passes are usually needed to find free blocks. Additionally, the
|
one or two passes are usually needed to find free blocks. Additionally, the
|
||||||
performance of the allocator can be optimized by adjusting the block size or
|
performance of the allocator can be optimized by adjusting the block size or
|
||||||
size of the lookahead buffer, trading either write granularity or RAM for
|
size of the lookahead buffer, trading either write granularity or RAM for
|
||||||
@@ -1173,9 +1173,9 @@ We may find that the new block is also bad, but hopefully after repeating this
|
|||||||
cycle we'll eventually find a new block where a write succeeds. If we don't,
|
cycle we'll eventually find a new block where a write succeeds. If we don't,
|
||||||
that means that all blocks in our storage are bad, and we've reached the end of
|
that means that all blocks in our storage are bad, and we've reached the end of
|
||||||
our device's usable life. At this point, littlefs will return an "out of space"
|
our device's usable life. At this point, littlefs will return an "out of space"
|
||||||
error, which is technically true, there are no more good blocks, but as an
|
error. This is technically true, as there are no more good blocks, but as an
|
||||||
added benefit also matches the error condition expected by users of dynamically
|
added benefit it also matches the error condition expected by users of
|
||||||
sized data.
|
dynamically sized data.
|
||||||
|
|
||||||
---
|
---
|
||||||
|
|
||||||
@@ -1187,7 +1187,7 @@ original data even after it has been corrupted. One such mechanism for this is
|
|||||||
ECC is an extension to the idea of a checksum. Where a checksum such as CRC can
|
ECC is an extension to the idea of a checksum. Where a checksum such as CRC can
|
||||||
detect that an error has occurred in the data, ECC can detect and actually
|
detect that an error has occurred in the data, ECC can detect and actually
|
||||||
correct some amount of errors. However, there is a limit to how many errors ECC
|
correct some amount of errors. However, there is a limit to how many errors ECC
|
||||||
can detect, call the [Hamming bound][wikipedia-hamming-bound]. As the number of
|
can detect: the [Hamming bound][wikipedia-hamming-bound]. As the number of
|
||||||
errors approaches the Hamming bound, we may still be able to detect errors, but
|
errors approaches the Hamming bound, we may still be able to detect errors, but
|
||||||
can no longer fix the data. If we've reached this point the block is
|
can no longer fix the data. If we've reached this point the block is
|
||||||
unrecoverable.
|
unrecoverable.
|
||||||
@@ -1202,7 +1202,7 @@ chip itself.
|
|||||||
In littlefs, ECC is entirely optional. Read errors can instead be prevented
|
In littlefs, ECC is entirely optional. Read errors can instead be prevented
|
||||||
proactively by wear leveling. But it's important to note that ECC can be used
|
proactively by wear leveling. But it's important to note that ECC can be used
|
||||||
at the block device level to modestly extend the life of a device. littlefs
|
at the block device level to modestly extend the life of a device. littlefs
|
||||||
respects any errors reported by the block device, allow a block device to
|
respects any errors reported by the block device, allowing a block device to
|
||||||
provide additional aggressive error detection.
|
provide additional aggressive error detection.
|
||||||
|
|
||||||
---
|
---
|
||||||
@@ -1231,7 +1231,7 @@ Generally, wear leveling algorithms fall into one of two categories:
|
|||||||
we need to consider all blocks, including blocks that already contain data.
|
we need to consider all blocks, including blocks that already contain data.
|
||||||
|
|
||||||
As a tradeoff for code size and complexity, littlefs (currently) only provides
|
As a tradeoff for code size and complexity, littlefs (currently) only provides
|
||||||
dynamic wear leveling. This is a best efforts solution. Wear is not distributed
|
dynamic wear leveling. This is a best effort solution. Wear is not distributed
|
||||||
perfectly, but it is distributed among the free blocks and greatly extends the
|
perfectly, but it is distributed among the free blocks and greatly extends the
|
||||||
life of a device.
|
life of a device.
|
||||||
|
|
||||||
@@ -1378,7 +1378,7 @@ We can make several improvements. First, instead of giving each file its own
|
|||||||
metadata pair, we can store multiple files in a single metadata pair. One way
|
metadata pair, we can store multiple files in a single metadata pair. One way
|
||||||
to do this is to directly associate a directory with a metadata pair (or a
|
to do this is to directly associate a directory with a metadata pair (or a
|
||||||
linked list of metadata pairs). This makes it easy for multiple files to share
|
linked list of metadata pairs). This makes it easy for multiple files to share
|
||||||
the directory's metadata pair for logging and reduce the collective storage
|
the directory's metadata pair for logging and reduces the collective storage
|
||||||
overhead.
|
overhead.
|
||||||
|
|
||||||
The strict binding of metadata pairs and directories also gives users
|
The strict binding of metadata pairs and directories also gives users
|
||||||
@@ -1816,12 +1816,12 @@ while manipulating the directory tree (foreshadowing!).
|
|||||||
|
|
||||||
## The move problem
|
## The move problem
|
||||||
|
|
||||||
We have one last challenge. The move problem. Phrasing the problem is simple:
|
We have one last challenge: the move problem. Phrasing the problem is simple:
|
||||||
|
|
||||||
How do you atomically move a file between two directories?
|
How do you atomically move a file between two directories?
|
||||||
|
|
||||||
In littlefs we can atomically commit to directories, but we can't create
|
In littlefs we can atomically commit to directories, but we can't create
|
||||||
an atomic commit that span multiple directories. The filesystem must go
|
an atomic commit that spans multiple directories. The filesystem must go
|
||||||
through a minimum of two distinct states to complete a move.
|
through a minimum of two distinct states to complete a move.
|
||||||
|
|
||||||
To make matters worse, file moves are a common form of synchronization for
|
To make matters worse, file moves are a common form of synchronization for
|
||||||
@@ -1831,13 +1831,13 @@ atomic moves right.
|
|||||||
So what can we do?
|
So what can we do?
|
||||||
|
|
||||||
- We definitely can't just let power-loss result in duplicated or lost files.
|
- We definitely can't just let power-loss result in duplicated or lost files.
|
||||||
This could easily break user's code and would only reveal itself in extreme
|
This could easily break users' code and would only reveal itself in extreme
|
||||||
cases. We were only able to be lazy about the threaded linked-list because
|
cases. We were only able to be lazy about the threaded linked-list because
|
||||||
it isn't user facing and we can handle the corner cases internally.
|
it isn't user facing and we can handle the corner cases internally.
|
||||||
|
|
||||||
- Some filesystems propagate COW operations up the tree until finding a common
|
- Some filesystems propagate COW operations up the tree until a common parent
|
||||||
parent. Unfortunately this interacts poorly with our threaded tree and brings
|
is found. Unfortunately this interacts poorly with our threaded tree and
|
||||||
back the issue of upward propagation of wear.
|
brings back the issue of upward propagation of wear.
|
||||||
|
|
||||||
- In a previous version of littlefs we tried to solve this problem by going
|
- In a previous version of littlefs we tried to solve this problem by going
|
||||||
back and forth between the source and destination, marking and unmarking the
|
back and forth between the source and destination, marking and unmarking the
|
||||||
@@ -1852,7 +1852,7 @@ introduction of a mechanism called "global state".
|
|||||||
---
|
---
|
||||||
|
|
||||||
Global state is a small set of state that can be updated from _any_ metadata
|
Global state is a small set of state that can be updated from _any_ metadata
|
||||||
pair. Combining global state with metadata pair's ability to update multiple
|
pair. Combining global state with metadata pairs' ability to update multiple
|
||||||
entries in one commit gives us a powerful tool for crafting complex atomic
|
entries in one commit gives us a powerful tool for crafting complex atomic
|
||||||
operations.
|
operations.
|
||||||
|
|
||||||
@@ -1910,7 +1910,7 @@ the filesystem is mounted.
|
|||||||
|
|
||||||
You may have noticed that global state is very expensive. We keep a copy in
|
You may have noticed that global state is very expensive. We keep a copy in
|
||||||
RAM and a delta in an unbounded number of metadata pairs. Even if we reset
|
RAM and a delta in an unbounded number of metadata pairs. Even if we reset
|
||||||
the global state to its initial value we can't easily clean up the deltas on
|
the global state to its initial value, we can't easily clean up the deltas on
|
||||||
disk. For this reason, it's very important that we keep the size of global
|
disk. For this reason, it's very important that we keep the size of global
|
||||||
state bounded and extremely small. But, even with a strict budget, global
|
state bounded and extremely small. But, even with a strict budget, global
|
||||||
state is incredibly valuable.
|
state is incredibly valuable.
|
||||||
|
|||||||
2
Makefile
2
Makefile
@@ -29,7 +29,7 @@ override CFLAGS += -DLFS_YES_TRACE
|
|||||||
endif
|
endif
|
||||||
override CFLAGS += -I.
|
override CFLAGS += -I.
|
||||||
override CFLAGS += -std=c99 -Wall -pedantic
|
override CFLAGS += -std=c99 -Wall -pedantic
|
||||||
override CFLAGS += -Wextra -Wshadow -Wjump-misses-init
|
override CFLAGS += -Wextra -Wshadow -Wjump-misses-init -Wundef
|
||||||
# Remove missing-field-initializers because of GCC bug
|
# Remove missing-field-initializers because of GCC bug
|
||||||
override CFLAGS += -Wno-missing-field-initializers
|
override CFLAGS += -Wno-missing-field-initializers
|
||||||
|
|
||||||
|
|||||||
@@ -53,6 +53,7 @@ const struct lfs_config cfg = {
|
|||||||
.block_count = 128,
|
.block_count = 128,
|
||||||
.cache_size = 16,
|
.cache_size = 16,
|
||||||
.lookahead_size = 16,
|
.lookahead_size = 16,
|
||||||
|
.block_cycles = 500,
|
||||||
};
|
};
|
||||||
|
|
||||||
// entry point
|
// entry point
|
||||||
|
|||||||
@@ -102,6 +102,7 @@ int lfs_emubd_create(const struct lfs_config *cfg, const char *path) {
|
|||||||
if (res < 1) {
|
if (res < 1) {
|
||||||
err = -errno;
|
err = -errno;
|
||||||
LFS_TRACE("lfs_emubd_create -> %"PRId32, err);
|
LFS_TRACE("lfs_emubd_create -> %"PRId32, err);
|
||||||
|
fclose(f);
|
||||||
return err;
|
return err;
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -124,6 +125,7 @@ int lfs_emubd_create(const struct lfs_config *cfg, const char *path) {
|
|||||||
if (res < 1) {
|
if (res < 1) {
|
||||||
err = -errno;
|
err = -errno;
|
||||||
LFS_TRACE("lfs_emubd_create -> %"PRId32, err);
|
LFS_TRACE("lfs_emubd_create -> %"PRId32, err);
|
||||||
|
fclose(f);
|
||||||
return err;
|
return err;
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -178,6 +180,7 @@ int lfs_emubd_read(const struct lfs_config *cfg, lfs_block_t block,
|
|||||||
if (err) {
|
if (err) {
|
||||||
err = -errno;
|
err = -errno;
|
||||||
LFS_TRACE("lfs_emubd_read -> %d", err);
|
LFS_TRACE("lfs_emubd_read -> %d", err);
|
||||||
|
fclose(f);
|
||||||
return err;
|
return err;
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -185,6 +188,7 @@ int lfs_emubd_read(const struct lfs_config *cfg, lfs_block_t block,
|
|||||||
if (res < size && !feof(f)) {
|
if (res < size && !feof(f)) {
|
||||||
err = -errno;
|
err = -errno;
|
||||||
LFS_TRACE("lfs_emubd_read -> %d", err);
|
LFS_TRACE("lfs_emubd_read -> %d", err);
|
||||||
|
fclose(f);
|
||||||
return err;
|
return err;
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -230,6 +234,7 @@ int lfs_emubd_prog(const struct lfs_config *cfg, lfs_block_t block,
|
|||||||
if (err) {
|
if (err) {
|
||||||
err = -errno;
|
err = -errno;
|
||||||
LFS_TRACE("lfs_emubd_prog -> %d", err);
|
LFS_TRACE("lfs_emubd_prog -> %d", err);
|
||||||
|
fclose(f);
|
||||||
return err;
|
return err;
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -237,6 +242,7 @@ int lfs_emubd_prog(const struct lfs_config *cfg, lfs_block_t block,
|
|||||||
if (res < size) {
|
if (res < size) {
|
||||||
err = -errno;
|
err = -errno;
|
||||||
LFS_TRACE("lfs_emubd_prog -> %d", err);
|
LFS_TRACE("lfs_emubd_prog -> %d", err);
|
||||||
|
fclose(f);
|
||||||
return err;
|
return err;
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -244,6 +250,7 @@ int lfs_emubd_prog(const struct lfs_config *cfg, lfs_block_t block,
|
|||||||
if (err) {
|
if (err) {
|
||||||
err = -errno;
|
err = -errno;
|
||||||
LFS_TRACE("lfs_emubd_prog -> %d", err);
|
LFS_TRACE("lfs_emubd_prog -> %d", err);
|
||||||
|
fclose(f);
|
||||||
return err;
|
return err;
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -252,6 +259,7 @@ int lfs_emubd_prog(const struct lfs_config *cfg, lfs_block_t block,
|
|||||||
if (res < 1) {
|
if (res < 1) {
|
||||||
err = -errno;
|
err = -errno;
|
||||||
LFS_TRACE("lfs_emubd_prog -> %d", err);
|
LFS_TRACE("lfs_emubd_prog -> %d", err);
|
||||||
|
fclose(f);
|
||||||
return err;
|
return err;
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -340,6 +348,7 @@ int lfs_emubd_sync(const struct lfs_config *cfg) {
|
|||||||
if (res < 1) {
|
if (res < 1) {
|
||||||
int err = -errno;
|
int err = -errno;
|
||||||
LFS_TRACE("lfs_emubd_sync -> %d", err);
|
LFS_TRACE("lfs_emubd_sync -> %d", err);
|
||||||
|
fclose(f);
|
||||||
return err;
|
return err;
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -364,6 +373,7 @@ int lfs_emubd_sync(const struct lfs_config *cfg) {
|
|||||||
if (res < 1) {
|
if (res < 1) {
|
||||||
err = -errno;
|
err = -errno;
|
||||||
LFS_TRACE("lfs_emubd_sync -> %d", err);
|
LFS_TRACE("lfs_emubd_sync -> %d", err);
|
||||||
|
fclose(f);
|
||||||
return err;
|
return err;
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -388,6 +398,7 @@ int lfs_emubd_sync(const struct lfs_config *cfg) {
|
|||||||
if (res < 1) {
|
if (res < 1) {
|
||||||
err = -errno;
|
err = -errno;
|
||||||
LFS_TRACE("lfs_emubd_sync -> %d", err);
|
LFS_TRACE("lfs_emubd_sync -> %d", err);
|
||||||
|
fclose(f);
|
||||||
return err;
|
return err;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
156
lfs.c
156
lfs.c
@@ -7,19 +7,21 @@
|
|||||||
#include "lfs.h"
|
#include "lfs.h"
|
||||||
#include "lfs_util.h"
|
#include "lfs_util.h"
|
||||||
|
|
||||||
|
#define LFS_BLOCK_NULL ((lfs_block_t)-1)
|
||||||
|
#define LFS_BLOCK_INLINE ((lfs_block_t)-2)
|
||||||
|
|
||||||
/// Caching block device operations ///
|
/// Caching block device operations ///
|
||||||
static inline void lfs_cache_drop(lfs_t *lfs, lfs_cache_t *rcache) {
|
static inline void lfs_cache_drop(lfs_t *lfs, lfs_cache_t *rcache) {
|
||||||
// do not zero, cheaper if cache is readonly or only going to be
|
// do not zero, cheaper if cache is readonly or only going to be
|
||||||
// written with identical data (during relocates)
|
// written with identical data (during relocates)
|
||||||
(void)lfs;
|
(void)lfs;
|
||||||
rcache->block = 0xffffffff;
|
rcache->block = LFS_BLOCK_NULL;
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline void lfs_cache_zero(lfs_t *lfs, lfs_cache_t *pcache) {
|
static inline void lfs_cache_zero(lfs_t *lfs, lfs_cache_t *pcache) {
|
||||||
// zero to avoid information leak
|
// zero to avoid information leak
|
||||||
memset(pcache->buffer, 0xff, lfs->cfg->cache_size);
|
memset(pcache->buffer, 0xff, lfs->cfg->cache_size);
|
||||||
pcache->block = 0xffffffff;
|
pcache->block = LFS_BLOCK_NULL;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int lfs_bd_read(lfs_t *lfs,
|
static int lfs_bd_read(lfs_t *lfs,
|
||||||
@@ -27,7 +29,7 @@ static int lfs_bd_read(lfs_t *lfs,
|
|||||||
lfs_block_t block, lfs_off_t off,
|
lfs_block_t block, lfs_off_t off,
|
||||||
void *buffer, lfs_size_t size) {
|
void *buffer, lfs_size_t size) {
|
||||||
uint8_t *data = buffer;
|
uint8_t *data = buffer;
|
||||||
LFS_ASSERT(block != 0xffffffff);
|
LFS_ASSERT(block != LFS_BLOCK_NULL);
|
||||||
if (off+size > lfs->cfg->block_size) {
|
if (off+size > lfs->cfg->block_size) {
|
||||||
return LFS_ERR_CORRUPT;
|
return LFS_ERR_CORRUPT;
|
||||||
}
|
}
|
||||||
@@ -121,7 +123,7 @@ static int lfs_bd_cmp(lfs_t *lfs,
|
|||||||
|
|
||||||
static int lfs_bd_flush(lfs_t *lfs,
|
static int lfs_bd_flush(lfs_t *lfs,
|
||||||
lfs_cache_t *pcache, lfs_cache_t *rcache, bool validate) {
|
lfs_cache_t *pcache, lfs_cache_t *rcache, bool validate) {
|
||||||
if (pcache->block != 0xffffffff && pcache->block != 0xfffffffe) {
|
if (pcache->block != LFS_BLOCK_NULL && pcache->block != LFS_BLOCK_INLINE) {
|
||||||
LFS_ASSERT(pcache->block < lfs->cfg->block_count);
|
LFS_ASSERT(pcache->block < lfs->cfg->block_count);
|
||||||
lfs_size_t diff = lfs_alignup(pcache->size, lfs->cfg->prog_size);
|
lfs_size_t diff = lfs_alignup(pcache->size, lfs->cfg->prog_size);
|
||||||
int err = lfs->cfg->prog(lfs->cfg, pcache->block,
|
int err = lfs->cfg->prog(lfs->cfg, pcache->block,
|
||||||
@@ -171,7 +173,7 @@ static int lfs_bd_prog(lfs_t *lfs,
|
|||||||
lfs_block_t block, lfs_off_t off,
|
lfs_block_t block, lfs_off_t off,
|
||||||
const void *buffer, lfs_size_t size) {
|
const void *buffer, lfs_size_t size) {
|
||||||
const uint8_t *data = buffer;
|
const uint8_t *data = buffer;
|
||||||
LFS_ASSERT(block != 0xffffffff);
|
LFS_ASSERT(block != LFS_BLOCK_NULL);
|
||||||
LFS_ASSERT(off + size <= lfs->cfg->block_size);
|
LFS_ASSERT(off + size <= lfs->cfg->block_size);
|
||||||
|
|
||||||
while (size > 0) {
|
while (size > 0) {
|
||||||
@@ -201,7 +203,7 @@ static int lfs_bd_prog(lfs_t *lfs,
|
|||||||
|
|
||||||
// pcache must have been flushed, either by programming and
|
// pcache must have been flushed, either by programming and
|
||||||
// entire block or manually flushing the pcache
|
// entire block or manually flushing the pcache
|
||||||
LFS_ASSERT(pcache->block == 0xffffffff);
|
LFS_ASSERT(pcache->block == LFS_BLOCK_NULL);
|
||||||
|
|
||||||
// prepare pcache, first condition can no longer fail
|
// prepare pcache, first condition can no longer fail
|
||||||
pcache->block = block;
|
pcache->block = block;
|
||||||
@@ -229,7 +231,7 @@ static inline void lfs_pair_swap(lfs_block_t pair[2]) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
static inline bool lfs_pair_isnull(const lfs_block_t pair[2]) {
|
static inline bool lfs_pair_isnull(const lfs_block_t pair[2]) {
|
||||||
return pair[0] == 0xffffffff || pair[1] == 0xffffffff;
|
return pair[0] == LFS_BLOCK_NULL || pair[1] == LFS_BLOCK_NULL;
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline int lfs_pair_cmp(
|
static inline int lfs_pair_cmp(
|
||||||
@@ -350,7 +352,7 @@ static inline bool lfs_gstate_hasmovehere(const struct lfs_gstate *a,
|
|||||||
|
|
||||||
static inline void lfs_gstate_xororphans(struct lfs_gstate *a,
|
static inline void lfs_gstate_xororphans(struct lfs_gstate *a,
|
||||||
const struct lfs_gstate *b, bool orphans) {
|
const struct lfs_gstate *b, bool orphans) {
|
||||||
a->tag ^= LFS_MKTAG(0x800, 0, 0) & (b->tag ^ (orphans << 31));
|
a->tag ^= LFS_MKTAG(0x800, 0, 0) & (b->tag ^ ((uint32_t)orphans << 31));
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline void lfs_gstate_xormove(struct lfs_gstate *a,
|
static inline void lfs_gstate_xormove(struct lfs_gstate *a,
|
||||||
@@ -571,7 +573,7 @@ static int lfs_dir_getread(lfs_t *lfs, const lfs_mdir_t *dir,
|
|||||||
while (size > 0) {
|
while (size > 0) {
|
||||||
lfs_size_t diff = size;
|
lfs_size_t diff = size;
|
||||||
|
|
||||||
if (pcache && pcache->block == 0xfffffffe &&
|
if (pcache && pcache->block == LFS_BLOCK_INLINE &&
|
||||||
off < pcache->off + pcache->size) {
|
off < pcache->off + pcache->size) {
|
||||||
if (off >= pcache->off) {
|
if (off >= pcache->off) {
|
||||||
// is already in pcache?
|
// is already in pcache?
|
||||||
@@ -588,7 +590,7 @@ static int lfs_dir_getread(lfs_t *lfs, const lfs_mdir_t *dir,
|
|||||||
diff = lfs_min(diff, pcache->off-off);
|
diff = lfs_min(diff, pcache->off-off);
|
||||||
}
|
}
|
||||||
|
|
||||||
if (rcache->block == 0xfffffffe &&
|
if (rcache->block == LFS_BLOCK_INLINE &&
|
||||||
off < rcache->off + rcache->size) {
|
off < rcache->off + rcache->size) {
|
||||||
if (off >= rcache->off) {
|
if (off >= rcache->off) {
|
||||||
// is already in rcache?
|
// is already in rcache?
|
||||||
@@ -606,7 +608,7 @@ static int lfs_dir_getread(lfs_t *lfs, const lfs_mdir_t *dir,
|
|||||||
}
|
}
|
||||||
|
|
||||||
// load to cache, first condition can no longer fail
|
// load to cache, first condition can no longer fail
|
||||||
rcache->block = 0xfffffffe;
|
rcache->block = LFS_BLOCK_INLINE;
|
||||||
rcache->off = lfs_aligndown(off, lfs->cfg->read_size);
|
rcache->off = lfs_aligndown(off, lfs->cfg->read_size);
|
||||||
rcache->size = lfs_min(lfs_alignup(off+hint, lfs->cfg->read_size),
|
rcache->size = lfs_min(lfs_alignup(off+hint, lfs->cfg->read_size),
|
||||||
lfs->cfg->cache_size);
|
lfs->cfg->cache_size);
|
||||||
@@ -723,7 +725,7 @@ static int lfs_dir_traverse(lfs_t *lfs,
|
|||||||
uint16_t fromid = lfs_tag_size(tag);
|
uint16_t fromid = lfs_tag_size(tag);
|
||||||
uint16_t toid = lfs_tag_id(tag);
|
uint16_t toid = lfs_tag_id(tag);
|
||||||
int err = lfs_dir_traverse(lfs,
|
int err = lfs_dir_traverse(lfs,
|
||||||
buffer, 0, 0xffffffff, NULL, 0, true,
|
buffer, 0, LFS_BLOCK_NULL, NULL, 0, true,
|
||||||
LFS_MKTAG(0x600, 0x3ff, 0),
|
LFS_MKTAG(0x600, 0x3ff, 0),
|
||||||
LFS_MKTAG(LFS_TYPE_STRUCT, 0, 0),
|
LFS_MKTAG(LFS_TYPE_STRUCT, 0, 0),
|
||||||
fromid, fromid+1, toid-fromid+diff,
|
fromid, fromid+1, toid-fromid+diff,
|
||||||
@@ -783,15 +785,15 @@ static lfs_stag_t lfs_dir_fetchmatch(lfs_t *lfs,
|
|||||||
// now scan tags to fetch the actual dir and find possible match
|
// now scan tags to fetch the actual dir and find possible match
|
||||||
for (int i = 0; i < 2; i++) {
|
for (int i = 0; i < 2; i++) {
|
||||||
lfs_off_t off = 0;
|
lfs_off_t off = 0;
|
||||||
lfs_tag_t ptag = 0xffffffff;
|
lfs_tag_t ptag = LFS_BLOCK_NULL;
|
||||||
|
|
||||||
uint16_t tempcount = 0;
|
uint16_t tempcount = 0;
|
||||||
lfs_block_t temptail[2] = {0xffffffff, 0xffffffff};
|
lfs_block_t temptail[2] = {LFS_BLOCK_NULL, LFS_BLOCK_NULL};
|
||||||
bool tempsplit = false;
|
bool tempsplit = false;
|
||||||
lfs_stag_t tempbesttag = besttag;
|
lfs_stag_t tempbesttag = besttag;
|
||||||
|
|
||||||
dir->rev = lfs_tole32(dir->rev);
|
dir->rev = lfs_tole32(dir->rev);
|
||||||
uint32_t crc = lfs_crc(0xffffffff, &dir->rev, sizeof(dir->rev));
|
uint32_t crc = lfs_crc(LFS_BLOCK_NULL, &dir->rev, sizeof(dir->rev));
|
||||||
dir->rev = lfs_fromle32(dir->rev);
|
dir->rev = lfs_fromle32(dir->rev);
|
||||||
|
|
||||||
while (true) {
|
while (true) {
|
||||||
@@ -844,7 +846,7 @@ static lfs_stag_t lfs_dir_fetchmatch(lfs_t *lfs,
|
|||||||
}
|
}
|
||||||
|
|
||||||
// reset the next bit if we need to
|
// reset the next bit if we need to
|
||||||
ptag ^= (lfs_tag_chunk(tag) & 1U) << 31;
|
ptag ^= (lfs_tag_t)(lfs_tag_chunk(tag) & 1U) << 31;
|
||||||
|
|
||||||
// toss our crc into the filesystem seed for
|
// toss our crc into the filesystem seed for
|
||||||
// pseudorandom numbers
|
// pseudorandom numbers
|
||||||
@@ -860,7 +862,7 @@ static lfs_stag_t lfs_dir_fetchmatch(lfs_t *lfs,
|
|||||||
dir->split = tempsplit;
|
dir->split = tempsplit;
|
||||||
|
|
||||||
// reset crc
|
// reset crc
|
||||||
crc = 0xffffffff;
|
crc = LFS_BLOCK_NULL;
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -975,7 +977,7 @@ static int lfs_dir_fetch(lfs_t *lfs,
|
|||||||
lfs_mdir_t *dir, const lfs_block_t pair[2]) {
|
lfs_mdir_t *dir, const lfs_block_t pair[2]) {
|
||||||
// note, mask=-1, tag=0 can never match a tag since this
|
// note, mask=-1, tag=0 can never match a tag since this
|
||||||
// pattern has the invalid bit set
|
// pattern has the invalid bit set
|
||||||
return lfs_dir_fetchmatch(lfs, dir, pair, -1, 0, NULL, NULL, NULL);
|
return (int)lfs_dir_fetchmatch(lfs, dir, pair, -1, 0, NULL, NULL, NULL);
|
||||||
}
|
}
|
||||||
|
|
||||||
static int lfs_dir_getgstate(lfs_t *lfs, const lfs_mdir_t *dir,
|
static int lfs_dir_getgstate(lfs_t *lfs, const lfs_mdir_t *dir,
|
||||||
@@ -1008,7 +1010,7 @@ static int lfs_dir_getinfo(lfs_t *lfs, lfs_mdir_t *dir,
|
|||||||
lfs_stag_t tag = lfs_dir_get(lfs, dir, LFS_MKTAG(0x780, 0x3ff, 0),
|
lfs_stag_t tag = lfs_dir_get(lfs, dir, LFS_MKTAG(0x780, 0x3ff, 0),
|
||||||
LFS_MKTAG(LFS_TYPE_NAME, id, lfs->name_max+1), info->name);
|
LFS_MKTAG(LFS_TYPE_NAME, id, lfs->name_max+1), info->name);
|
||||||
if (tag < 0) {
|
if (tag < 0) {
|
||||||
return tag;
|
return (int)tag;
|
||||||
}
|
}
|
||||||
|
|
||||||
info->type = lfs_tag_type3(tag);
|
info->type = lfs_tag_type3(tag);
|
||||||
@@ -1017,7 +1019,7 @@ static int lfs_dir_getinfo(lfs_t *lfs, lfs_mdir_t *dir,
|
|||||||
tag = lfs_dir_get(lfs, dir, LFS_MKTAG(0x700, 0x3ff, 0),
|
tag = lfs_dir_get(lfs, dir, LFS_MKTAG(0x700, 0x3ff, 0),
|
||||||
LFS_MKTAG(LFS_TYPE_STRUCT, id, sizeof(ctz)), &ctz);
|
LFS_MKTAG(LFS_TYPE_STRUCT, id, sizeof(ctz)), &ctz);
|
||||||
if (tag < 0) {
|
if (tag < 0) {
|
||||||
return tag;
|
return (int)tag;
|
||||||
}
|
}
|
||||||
lfs_ctz_fromle32(&ctz);
|
lfs_ctz_fromle32(&ctz);
|
||||||
|
|
||||||
@@ -1060,7 +1062,7 @@ static int lfs_dir_find_match(void *data,
|
|||||||
return LFS_CMP_EQ;
|
return LFS_CMP_EQ;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int lfs_dir_find(lfs_t *lfs, lfs_mdir_t *dir,
|
static lfs_stag_t lfs_dir_find(lfs_t *lfs, lfs_mdir_t *dir,
|
||||||
const char **path, uint16_t *id) {
|
const char **path, uint16_t *id) {
|
||||||
// we reduce path to a single name if we can find it
|
// we reduce path to a single name if we can find it
|
||||||
const char *name = *path;
|
const char *name = *path;
|
||||||
@@ -1248,7 +1250,7 @@ static int lfs_dir_commitcrc(lfs_t *lfs, struct lfs_commit *commit) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// read erased state from next program unit
|
// read erased state from next program unit
|
||||||
lfs_tag_t tag = 0xffffffff;
|
lfs_tag_t tag = LFS_BLOCK_NULL;
|
||||||
int err = lfs_bd_read(lfs,
|
int err = lfs_bd_read(lfs,
|
||||||
NULL, &lfs->rcache, sizeof(tag),
|
NULL, &lfs->rcache, sizeof(tag),
|
||||||
commit->block, noff, &tag, sizeof(tag));
|
commit->block, noff, &tag, sizeof(tag));
|
||||||
@@ -1273,8 +1275,8 @@ static int lfs_dir_commitcrc(lfs_t *lfs, struct lfs_commit *commit) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
commit->off += sizeof(tag)+lfs_tag_size(tag);
|
commit->off += sizeof(tag)+lfs_tag_size(tag);
|
||||||
commit->ptag = tag ^ (reset << 31);
|
commit->ptag = tag ^ ((lfs_tag_t)reset << 31);
|
||||||
commit->crc = 0xffffffff; // reset crc for next "commit"
|
commit->crc = LFS_BLOCK_NULL; // reset crc for next "commit"
|
||||||
}
|
}
|
||||||
|
|
||||||
// flush buffers
|
// flush buffers
|
||||||
@@ -1287,7 +1289,7 @@ static int lfs_dir_commitcrc(lfs_t *lfs, struct lfs_commit *commit) {
|
|||||||
lfs_off_t off = commit->begin;
|
lfs_off_t off = commit->begin;
|
||||||
lfs_off_t noff = off1;
|
lfs_off_t noff = off1;
|
||||||
while (off < end) {
|
while (off < end) {
|
||||||
uint32_t crc = 0xffffffff;
|
uint32_t crc = LFS_BLOCK_NULL;
|
||||||
for (lfs_off_t i = off; i < noff+sizeof(uint32_t); i++) {
|
for (lfs_off_t i = off; i < noff+sizeof(uint32_t); i++) {
|
||||||
// leave it up to caching to make this efficient
|
// leave it up to caching to make this efficient
|
||||||
uint8_t dat;
|
uint8_t dat;
|
||||||
@@ -1341,10 +1343,10 @@ static int lfs_dir_alloc(lfs_t *lfs, lfs_mdir_t *dir) {
|
|||||||
|
|
||||||
// set defaults
|
// set defaults
|
||||||
dir->off = sizeof(dir->rev);
|
dir->off = sizeof(dir->rev);
|
||||||
dir->etag = 0xffffffff;
|
dir->etag = LFS_BLOCK_NULL;
|
||||||
dir->count = 0;
|
dir->count = 0;
|
||||||
dir->tail[0] = 0xffffffff;
|
dir->tail[0] = LFS_BLOCK_NULL;
|
||||||
dir->tail[1] = 0xffffffff;
|
dir->tail[1] = LFS_BLOCK_NULL;
|
||||||
dir->erased = false;
|
dir->erased = false;
|
||||||
dir->split = false;
|
dir->split = false;
|
||||||
|
|
||||||
@@ -1434,7 +1436,7 @@ static int lfs_dir_compact(lfs_t *lfs,
|
|||||||
// find size
|
// find size
|
||||||
lfs_size_t size = 0;
|
lfs_size_t size = 0;
|
||||||
int err = lfs_dir_traverse(lfs,
|
int err = lfs_dir_traverse(lfs,
|
||||||
source, 0, 0xffffffff, attrs, attrcount, false,
|
source, 0, LFS_BLOCK_NULL, attrs, attrcount, false,
|
||||||
LFS_MKTAG(0x400, 0x3ff, 0),
|
LFS_MKTAG(0x400, 0x3ff, 0),
|
||||||
LFS_MKTAG(LFS_TYPE_NAME, 0, 0),
|
LFS_MKTAG(LFS_TYPE_NAME, 0, 0),
|
||||||
begin, end, -begin,
|
begin, end, -begin,
|
||||||
@@ -1500,7 +1502,7 @@ static int lfs_dir_compact(lfs_t *lfs,
|
|||||||
end = begin;
|
end = begin;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
#if LFS_MIGRATE
|
#ifdef LFS_MIGRATE
|
||||||
} else if (lfs_pair_cmp(dir->pair, lfs->root) == 0 && lfs->lfs1) {
|
} else if (lfs_pair_cmp(dir->pair, lfs->root) == 0 && lfs->lfs1) {
|
||||||
// we can't relocate our root during migrations, as this would
|
// we can't relocate our root during migrations, as this would
|
||||||
// cause the superblock to get updated, which would clobber v1
|
// cause the superblock to get updated, which would clobber v1
|
||||||
@@ -1526,8 +1528,8 @@ static int lfs_dir_compact(lfs_t *lfs,
|
|||||||
struct lfs_commit commit = {
|
struct lfs_commit commit = {
|
||||||
.block = dir->pair[1],
|
.block = dir->pair[1],
|
||||||
.off = 0,
|
.off = 0,
|
||||||
.ptag = 0xffffffff,
|
.ptag = LFS_BLOCK_NULL,
|
||||||
.crc = 0xffffffff,
|
.crc = LFS_BLOCK_NULL,
|
||||||
|
|
||||||
.begin = 0,
|
.begin = 0,
|
||||||
.end = lfs->cfg->block_size - 8,
|
.end = lfs->cfg->block_size - 8,
|
||||||
@@ -1556,7 +1558,7 @@ static int lfs_dir_compact(lfs_t *lfs,
|
|||||||
|
|
||||||
// traverse the directory, this time writing out all unique tags
|
// traverse the directory, this time writing out all unique tags
|
||||||
err = lfs_dir_traverse(lfs,
|
err = lfs_dir_traverse(lfs,
|
||||||
source, 0, 0xffffffff, attrs, attrcount, false,
|
source, 0, LFS_BLOCK_NULL, attrs, attrcount, false,
|
||||||
LFS_MKTAG(0x400, 0x3ff, 0),
|
LFS_MKTAG(0x400, 0x3ff, 0),
|
||||||
LFS_MKTAG(LFS_TYPE_NAME, 0, 0),
|
LFS_MKTAG(LFS_TYPE_NAME, 0, 0),
|
||||||
begin, end, -begin,
|
begin, end, -begin,
|
||||||
@@ -1682,8 +1684,8 @@ static int lfs_dir_commit(lfs_t *lfs, lfs_mdir_t *dir,
|
|||||||
}
|
}
|
||||||
|
|
||||||
// calculate changes to the directory
|
// calculate changes to the directory
|
||||||
lfs_tag_t deletetag = 0xffffffff;
|
lfs_tag_t deletetag = LFS_BLOCK_NULL;
|
||||||
lfs_tag_t createtag = 0xffffffff;
|
lfs_tag_t createtag = LFS_BLOCK_NULL;
|
||||||
for (int i = 0; i < attrcount; i++) {
|
for (int i = 0; i < attrcount; i++) {
|
||||||
if (lfs_tag_type3(attrs[i].tag) == LFS_TYPE_CREATE) {
|
if (lfs_tag_type3(attrs[i].tag) == LFS_TYPE_CREATE) {
|
||||||
createtag = attrs[i].tag;
|
createtag = attrs[i].tag;
|
||||||
@@ -1729,7 +1731,7 @@ static int lfs_dir_commit(lfs_t *lfs, lfs_mdir_t *dir,
|
|||||||
.block = dir->pair[0],
|
.block = dir->pair[0],
|
||||||
.off = dir->off,
|
.off = dir->off,
|
||||||
.ptag = dir->etag,
|
.ptag = dir->etag,
|
||||||
.crc = 0xffffffff,
|
.crc = LFS_BLOCK_NULL,
|
||||||
|
|
||||||
.begin = dir->off,
|
.begin = dir->off,
|
||||||
.end = lfs->cfg->block_size - 8,
|
.end = lfs->cfg->block_size - 8,
|
||||||
@@ -1813,8 +1815,8 @@ compact:
|
|||||||
if (lfs_pair_cmp(d->m.pair, copy.pair) == 0) {
|
if (lfs_pair_cmp(d->m.pair, copy.pair) == 0) {
|
||||||
d->m = *dir;
|
d->m = *dir;
|
||||||
if (d->id == lfs_tag_id(deletetag)) {
|
if (d->id == lfs_tag_id(deletetag)) {
|
||||||
d->m.pair[0] = 0xffffffff;
|
d->m.pair[0] = LFS_BLOCK_NULL;
|
||||||
d->m.pair[1] = 0xffffffff;
|
d->m.pair[1] = LFS_BLOCK_NULL;
|
||||||
} else if (d->id > lfs_tag_id(deletetag)) {
|
} else if (d->id > lfs_tag_id(deletetag)) {
|
||||||
d->id -= 1;
|
d->id -= 1;
|
||||||
if (d->type == LFS_TYPE_DIR) {
|
if (d->type == LFS_TYPE_DIR) {
|
||||||
@@ -2129,7 +2131,7 @@ static int lfs_ctz_find(lfs_t *lfs,
|
|||||||
lfs_block_t head, lfs_size_t size,
|
lfs_block_t head, lfs_size_t size,
|
||||||
lfs_size_t pos, lfs_block_t *block, lfs_off_t *off) {
|
lfs_size_t pos, lfs_block_t *block, lfs_off_t *off) {
|
||||||
if (size == 0) {
|
if (size == 0) {
|
||||||
*block = 0xffffffff;
|
*block = LFS_BLOCK_NULL;
|
||||||
*off = 0;
|
*off = 0;
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
@@ -2327,6 +2329,7 @@ int lfs_file_opencfg(lfs_t *lfs, lfs_file_t *file,
|
|||||||
file->cfg = cfg;
|
file->cfg = cfg;
|
||||||
file->flags = flags | LFS_F_OPENED;
|
file->flags = flags | LFS_F_OPENED;
|
||||||
file->pos = 0;
|
file->pos = 0;
|
||||||
|
file->off = 0;
|
||||||
file->cache.buffer = NULL;
|
file->cache.buffer = NULL;
|
||||||
|
|
||||||
// allocate entry for file if it doesn't exist
|
// allocate entry for file if it doesn't exist
|
||||||
@@ -2426,7 +2429,7 @@ int lfs_file_opencfg(lfs_t *lfs, lfs_file_t *file,
|
|||||||
|
|
||||||
if (lfs_tag_type3(tag) == LFS_TYPE_INLINESTRUCT) {
|
if (lfs_tag_type3(tag) == LFS_TYPE_INLINESTRUCT) {
|
||||||
// load inline files
|
// load inline files
|
||||||
file->ctz.head = 0xfffffffe;
|
file->ctz.head = LFS_BLOCK_INLINE;
|
||||||
file->ctz.size = lfs_tag_size(tag);
|
file->ctz.size = lfs_tag_size(tag);
|
||||||
file->flags |= LFS_F_INLINE;
|
file->flags |= LFS_F_INLINE;
|
||||||
file->cache.block = file->ctz.head;
|
file->cache.block = file->ctz.head;
|
||||||
@@ -2614,7 +2617,7 @@ static int lfs_file_flush(lfs_t *lfs, lfs_file_t *file) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// keep our reference to the rcache in sync
|
// keep our reference to the rcache in sync
|
||||||
if (lfs->rcache.block != 0xffffffff) {
|
if (lfs->rcache.block != LFS_BLOCK_NULL) {
|
||||||
lfs_cache_drop(lfs, &orig.cache);
|
lfs_cache_drop(lfs, &orig.cache);
|
||||||
lfs_cache_drop(lfs, &lfs->rcache);
|
lfs_cache_drop(lfs, &lfs->rcache);
|
||||||
}
|
}
|
||||||
@@ -2762,7 +2765,7 @@ lfs_ssize_t lfs_file_read(lfs_t *lfs, lfs_file_t *file,
|
|||||||
return err;
|
return err;
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
file->block = 0xfffffffe;
|
file->block = LFS_BLOCK_INLINE;
|
||||||
file->off = file->pos;
|
file->off = file->pos;
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -2888,7 +2891,7 @@ lfs_ssize_t lfs_file_write(lfs_t *lfs, lfs_file_t *file,
|
|||||||
return err;
|
return err;
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
file->block = 0xfffffffe;
|
file->block = LFS_BLOCK_INLINE;
|
||||||
file->off = file->pos;
|
file->off = file->pos;
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -2978,6 +2981,7 @@ int lfs_file_truncate(lfs_t *lfs, lfs_file_t *file, lfs_off_t size) {
|
|||||||
return LFS_ERR_INVAL;
|
return LFS_ERR_INVAL;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
lfs_off_t pos = file->pos;
|
||||||
lfs_off_t oldsize = lfs_file_size(lfs, file);
|
lfs_off_t oldsize = lfs_file_size(lfs, file);
|
||||||
if (size < oldsize) {
|
if (size < oldsize) {
|
||||||
// need to flush since directly changing metadata
|
// need to flush since directly changing metadata
|
||||||
@@ -3000,14 +3004,12 @@ int lfs_file_truncate(lfs_t *lfs, lfs_file_t *file, lfs_off_t size) {
|
|||||||
file->ctz.size = size;
|
file->ctz.size = size;
|
||||||
file->flags |= LFS_F_DIRTY | LFS_F_READING;
|
file->flags |= LFS_F_DIRTY | LFS_F_READING;
|
||||||
} else if (size > oldsize) {
|
} else if (size > oldsize) {
|
||||||
lfs_off_t pos = file->pos;
|
|
||||||
|
|
||||||
// flush+seek if not already at end
|
// flush+seek if not already at end
|
||||||
if (file->pos != oldsize) {
|
if (file->pos != oldsize) {
|
||||||
int err = lfs_file_seek(lfs, file, 0, LFS_SEEK_END);
|
lfs_soff_t res = lfs_file_seek(lfs, file, 0, LFS_SEEK_END);
|
||||||
if (err < 0) {
|
if (res < 0) {
|
||||||
LFS_TRACE("lfs_file_truncate -> %d", err);
|
LFS_TRACE("lfs_file_truncate -> %d", res);
|
||||||
return err;
|
return (int)res;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -3016,16 +3018,16 @@ int lfs_file_truncate(lfs_t *lfs, lfs_file_t *file, lfs_off_t size) {
|
|||||||
lfs_ssize_t res = lfs_file_write(lfs, file, &(uint8_t){0}, 1);
|
lfs_ssize_t res = lfs_file_write(lfs, file, &(uint8_t){0}, 1);
|
||||||
if (res < 0) {
|
if (res < 0) {
|
||||||
LFS_TRACE("lfs_file_truncate -> %d", res);
|
LFS_TRACE("lfs_file_truncate -> %d", res);
|
||||||
return res;
|
return (int)res;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
}
|
||||||
|
|
||||||
// restore pos
|
// restore pos
|
||||||
int err = lfs_file_seek(lfs, file, pos, LFS_SEEK_SET);
|
lfs_soff_t res = lfs_file_seek(lfs, file, pos, LFS_SEEK_SET);
|
||||||
if (err < 0) {
|
if (res < 0) {
|
||||||
LFS_TRACE("lfs_file_truncate -> %d", err);
|
LFS_TRACE("lfs_file_truncate -> %d", res);
|
||||||
return err;
|
return (int)res;
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
LFS_TRACE("lfs_file_truncate -> %d", 0);
|
LFS_TRACE("lfs_file_truncate -> %d", 0);
|
||||||
@@ -3045,7 +3047,7 @@ int lfs_file_rewind(lfs_t *lfs, lfs_file_t *file) {
|
|||||||
lfs_soff_t res = lfs_file_seek(lfs, file, 0, LFS_SEEK_SET);
|
lfs_soff_t res = lfs_file_seek(lfs, file, 0, LFS_SEEK_SET);
|
||||||
if (res < 0) {
|
if (res < 0) {
|
||||||
LFS_TRACE("lfs_file_rewind -> %d", res);
|
LFS_TRACE("lfs_file_rewind -> %d", res);
|
||||||
return res;
|
return (int)res;
|
||||||
}
|
}
|
||||||
|
|
||||||
LFS_TRACE("lfs_file_rewind -> %d", 0);
|
LFS_TRACE("lfs_file_rewind -> %d", 0);
|
||||||
@@ -3074,7 +3076,7 @@ int lfs_stat(lfs_t *lfs, const char *path, struct lfs_info *info) {
|
|||||||
lfs_stag_t tag = lfs_dir_find(lfs, &cwd, &path, NULL);
|
lfs_stag_t tag = lfs_dir_find(lfs, &cwd, &path, NULL);
|
||||||
if (tag < 0) {
|
if (tag < 0) {
|
||||||
LFS_TRACE("lfs_stat -> %d", tag);
|
LFS_TRACE("lfs_stat -> %d", tag);
|
||||||
return tag;
|
return (int)tag;
|
||||||
}
|
}
|
||||||
|
|
||||||
int err = lfs_dir_getinfo(lfs, &cwd, lfs_tag_id(tag), info);
|
int err = lfs_dir_getinfo(lfs, &cwd, lfs_tag_id(tag), info);
|
||||||
@@ -3095,7 +3097,7 @@ int lfs_remove(lfs_t *lfs, const char *path) {
|
|||||||
lfs_stag_t tag = lfs_dir_find(lfs, &cwd, &path, NULL);
|
lfs_stag_t tag = lfs_dir_find(lfs, &cwd, &path, NULL);
|
||||||
if (tag < 0 || lfs_tag_id(tag) == 0x3ff) {
|
if (tag < 0 || lfs_tag_id(tag) == 0x3ff) {
|
||||||
LFS_TRACE("lfs_remove -> %d", (tag < 0) ? tag : LFS_ERR_INVAL);
|
LFS_TRACE("lfs_remove -> %d", (tag < 0) ? tag : LFS_ERR_INVAL);
|
||||||
return (tag < 0) ? tag : LFS_ERR_INVAL;
|
return (tag < 0) ? (int)tag : LFS_ERR_INVAL;
|
||||||
}
|
}
|
||||||
|
|
||||||
lfs_mdir_t dir;
|
lfs_mdir_t dir;
|
||||||
@@ -3106,7 +3108,7 @@ int lfs_remove(lfs_t *lfs, const char *path) {
|
|||||||
LFS_MKTAG(LFS_TYPE_STRUCT, lfs_tag_id(tag), 8), pair);
|
LFS_MKTAG(LFS_TYPE_STRUCT, lfs_tag_id(tag), 8), pair);
|
||||||
if (res < 0) {
|
if (res < 0) {
|
||||||
LFS_TRACE("lfs_remove -> %d", res);
|
LFS_TRACE("lfs_remove -> %d", res);
|
||||||
return res;
|
return (int)res;
|
||||||
}
|
}
|
||||||
lfs_pair_fromle32(pair);
|
lfs_pair_fromle32(pair);
|
||||||
|
|
||||||
@@ -3169,7 +3171,7 @@ int lfs_rename(lfs_t *lfs, const char *oldpath, const char *newpath) {
|
|||||||
lfs_stag_t oldtag = lfs_dir_find(lfs, &oldcwd, &oldpath, NULL);
|
lfs_stag_t oldtag = lfs_dir_find(lfs, &oldcwd, &oldpath, NULL);
|
||||||
if (oldtag < 0 || lfs_tag_id(oldtag) == 0x3ff) {
|
if (oldtag < 0 || lfs_tag_id(oldtag) == 0x3ff) {
|
||||||
LFS_TRACE("lfs_rename -> %d", (oldtag < 0) ? oldtag : LFS_ERR_INVAL);
|
LFS_TRACE("lfs_rename -> %d", (oldtag < 0) ? oldtag : LFS_ERR_INVAL);
|
||||||
return (oldtag < 0) ? oldtag : LFS_ERR_INVAL;
|
return (oldtag < 0) ? (int)oldtag : LFS_ERR_INVAL;
|
||||||
}
|
}
|
||||||
|
|
||||||
// find new entry
|
// find new entry
|
||||||
@@ -3179,7 +3181,7 @@ int lfs_rename(lfs_t *lfs, const char *oldpath, const char *newpath) {
|
|||||||
if ((prevtag < 0 || lfs_tag_id(prevtag) == 0x3ff) &&
|
if ((prevtag < 0 || lfs_tag_id(prevtag) == 0x3ff) &&
|
||||||
!(prevtag == LFS_ERR_NOENT && newid != 0x3ff)) {
|
!(prevtag == LFS_ERR_NOENT && newid != 0x3ff)) {
|
||||||
LFS_TRACE("lfs_rename -> %d", (prevtag < 0) ? prevtag : LFS_ERR_INVAL);
|
LFS_TRACE("lfs_rename -> %d", (prevtag < 0) ? prevtag : LFS_ERR_INVAL);
|
||||||
return (prevtag < 0) ? prevtag : LFS_ERR_INVAL;
|
return (prevtag < 0) ? (int)prevtag : LFS_ERR_INVAL;
|
||||||
}
|
}
|
||||||
|
|
||||||
lfs_mdir_t prevdir;
|
lfs_mdir_t prevdir;
|
||||||
@@ -3200,7 +3202,7 @@ int lfs_rename(lfs_t *lfs, const char *oldpath, const char *newpath) {
|
|||||||
LFS_MKTAG(LFS_TYPE_STRUCT, newid, 8), prevpair);
|
LFS_MKTAG(LFS_TYPE_STRUCT, newid, 8), prevpair);
|
||||||
if (res < 0) {
|
if (res < 0) {
|
||||||
LFS_TRACE("lfs_rename -> %d", res);
|
LFS_TRACE("lfs_rename -> %d", res);
|
||||||
return res;
|
return (int)res;
|
||||||
}
|
}
|
||||||
lfs_pair_fromle32(prevpair);
|
lfs_pair_fromle32(prevpair);
|
||||||
|
|
||||||
@@ -3367,6 +3369,12 @@ static int lfs_init(lfs_t *lfs, const struct lfs_config *cfg) {
|
|||||||
lfs->cfg = cfg;
|
lfs->cfg = cfg;
|
||||||
int err = 0;
|
int err = 0;
|
||||||
|
|
||||||
|
// validate that the lfs-cfg sizes were initiated properly before
|
||||||
|
// performing any arithmetic logics with them
|
||||||
|
LFS_ASSERT(lfs->cfg->read_size != 0);
|
||||||
|
LFS_ASSERT(lfs->cfg->prog_size != 0);
|
||||||
|
LFS_ASSERT(lfs->cfg->cache_size != 0);
|
||||||
|
|
||||||
// check that block size is a multiple of cache size is a multiple
|
// check that block size is a multiple of cache size is a multiple
|
||||||
// of prog and read sizes
|
// of prog and read sizes
|
||||||
LFS_ASSERT(lfs->cfg->cache_size % lfs->cfg->read_size == 0);
|
LFS_ASSERT(lfs->cfg->cache_size % lfs->cfg->read_size == 0);
|
||||||
@@ -3374,7 +3382,7 @@ static int lfs_init(lfs_t *lfs, const struct lfs_config *cfg) {
|
|||||||
LFS_ASSERT(lfs->cfg->block_size % lfs->cfg->cache_size == 0);
|
LFS_ASSERT(lfs->cfg->block_size % lfs->cfg->cache_size == 0);
|
||||||
|
|
||||||
// check that the block size is large enough to fit ctz pointers
|
// check that the block size is large enough to fit ctz pointers
|
||||||
LFS_ASSERT(4*lfs_npw2(0xffffffff / (lfs->cfg->block_size-2*4))
|
LFS_ASSERT(4*lfs_npw2(LFS_BLOCK_NULL / (lfs->cfg->block_size-2*4))
|
||||||
<= lfs->cfg->block_size);
|
<= lfs->cfg->block_size);
|
||||||
|
|
||||||
// block_cycles = 0 is no longer supported.
|
// block_cycles = 0 is no longer supported.
|
||||||
@@ -3412,10 +3420,10 @@ static int lfs_init(lfs_t *lfs, const struct lfs_config *cfg) {
|
|||||||
lfs_cache_zero(lfs, &lfs->rcache);
|
lfs_cache_zero(lfs, &lfs->rcache);
|
||||||
lfs_cache_zero(lfs, &lfs->pcache);
|
lfs_cache_zero(lfs, &lfs->pcache);
|
||||||
|
|
||||||
// setup lookahead, must be multiple of 64-bits
|
// setup lookahead, must be multiple of 64-bits, 32-bit aligned
|
||||||
LFS_ASSERT(lfs->cfg->lookahead_size > 0);
|
LFS_ASSERT(lfs->cfg->lookahead_size > 0);
|
||||||
LFS_ASSERT(lfs->cfg->lookahead_size % 8 == 0 &&
|
LFS_ASSERT(lfs->cfg->lookahead_size % 8 == 0 &&
|
||||||
(uintptr_t)lfs->cfg->lookahead_buffer % 8 == 0);
|
(uintptr_t)lfs->cfg->lookahead_buffer % 4 == 0);
|
||||||
if (lfs->cfg->lookahead_buffer) {
|
if (lfs->cfg->lookahead_buffer) {
|
||||||
lfs->free.buffer = lfs->cfg->lookahead_buffer;
|
lfs->free.buffer = lfs->cfg->lookahead_buffer;
|
||||||
} else {
|
} else {
|
||||||
@@ -3446,8 +3454,8 @@ static int lfs_init(lfs_t *lfs, const struct lfs_config *cfg) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// setup default state
|
// setup default state
|
||||||
lfs->root[0] = 0xffffffff;
|
lfs->root[0] = LFS_BLOCK_NULL;
|
||||||
lfs->root[1] = 0xffffffff;
|
lfs->root[1] = LFS_BLOCK_NULL;
|
||||||
lfs->mlist = NULL;
|
lfs->mlist = NULL;
|
||||||
lfs->seed = 0;
|
lfs->seed = 0;
|
||||||
lfs->gstate = (struct lfs_gstate){0};
|
lfs->gstate = (struct lfs_gstate){0};
|
||||||
@@ -4250,7 +4258,7 @@ static int lfs1_dir_fetch(lfs_t *lfs,
|
|||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
|
|
||||||
uint32_t crc = 0xffffffff;
|
uint32_t crc = LFS_BLOCK_NULL;
|
||||||
lfs1_dir_tole32(&test);
|
lfs1_dir_tole32(&test);
|
||||||
lfs1_crc(&crc, &test, sizeof(test));
|
lfs1_crc(&crc, &test, sizeof(test));
|
||||||
lfs1_dir_fromle32(&test);
|
lfs1_dir_fromle32(&test);
|
||||||
@@ -4435,8 +4443,8 @@ static int lfs1_mount(lfs_t *lfs, struct lfs1 *lfs1,
|
|||||||
}
|
}
|
||||||
|
|
||||||
lfs->lfs1 = lfs1;
|
lfs->lfs1 = lfs1;
|
||||||
lfs->lfs1->root[0] = 0xffffffff;
|
lfs->lfs1->root[0] = LFS_BLOCK_NULL;
|
||||||
lfs->lfs1->root[1] = 0xffffffff;
|
lfs->lfs1->root[1] = LFS_BLOCK_NULL;
|
||||||
|
|
||||||
// setup free lookahead
|
// setup free lookahead
|
||||||
lfs->free.off = 0;
|
lfs->free.off = 0;
|
||||||
@@ -4606,7 +4614,7 @@ int lfs_migrate(lfs_t *lfs, const struct lfs_config *cfg) {
|
|||||||
id, entry1.d.nlen), name},
|
id, entry1.d.nlen), name},
|
||||||
{LFS_MKTAG(
|
{LFS_MKTAG(
|
||||||
isdir ? LFS_TYPE_DIRSTRUCT : LFS_TYPE_CTZSTRUCT,
|
isdir ? LFS_TYPE_DIRSTRUCT : LFS_TYPE_CTZSTRUCT,
|
||||||
id, sizeof(&entry1.d.u)), &entry1.d.u}));
|
id, sizeof(entry1.d.u)), &entry1.d.u}));
|
||||||
lfs1_entry_fromle32(&entry1.d);
|
lfs1_entry_fromle32(&entry1.d);
|
||||||
if (err) {
|
if (err) {
|
||||||
goto cleanup;
|
goto cleanup;
|
||||||
@@ -4629,7 +4637,7 @@ int lfs_migrate(lfs_t *lfs, const struct lfs_config *cfg) {
|
|||||||
|
|
||||||
lfs_pair_tole32(dir2.pair);
|
lfs_pair_tole32(dir2.pair);
|
||||||
err = lfs_dir_commit(lfs, &dir2, LFS_MKATTRS(
|
err = lfs_dir_commit(lfs, &dir2, LFS_MKATTRS(
|
||||||
{LFS_MKTAG(LFS_TYPE_SOFTTAIL, 0x3ff, 0),
|
{LFS_MKTAG(LFS_TYPE_SOFTTAIL, 0x3ff, 8),
|
||||||
dir1.d.tail}));
|
dir1.d.tail}));
|
||||||
lfs_pair_fromle32(dir2.pair);
|
lfs_pair_fromle32(dir2.pair);
|
||||||
if (err) {
|
if (err) {
|
||||||
@@ -4685,7 +4693,7 @@ int lfs_migrate(lfs_t *lfs, const struct lfs_config *cfg) {
|
|||||||
dir2.pair[1] = dir1.pair[1];
|
dir2.pair[1] = dir1.pair[1];
|
||||||
dir2.rev = dir1.d.rev;
|
dir2.rev = dir1.d.rev;
|
||||||
dir2.off = sizeof(dir2.rev);
|
dir2.off = sizeof(dir2.rev);
|
||||||
dir2.etag = 0xffffffff;
|
dir2.etag = LFS_BLOCK_NULL;
|
||||||
dir2.count = 0;
|
dir2.count = 0;
|
||||||
dir2.tail[0] = lfs->lfs1->root[0];
|
dir2.tail[0] = lfs->lfs1->root[0];
|
||||||
dir2.tail[1] = lfs->lfs1->root[1];
|
dir2.tail[1] = lfs->lfs1->root[1];
|
||||||
|
|||||||
@@ -179,7 +179,7 @@ echo "--- Really big path test ---"
|
|||||||
scripts/test.py << TEST
|
scripts/test.py << TEST
|
||||||
lfs_mount(&lfs, &cfg) => 0;
|
lfs_mount(&lfs, &cfg) => 0;
|
||||||
memset(path, 'w', LFS_NAME_MAX);
|
memset(path, 'w', LFS_NAME_MAX);
|
||||||
path[LFS_NAME_MAX+1] = '\0';
|
path[LFS_NAME_MAX] = '\0';
|
||||||
lfs_mkdir(&lfs, path) => 0;
|
lfs_mkdir(&lfs, path) => 0;
|
||||||
lfs_remove(&lfs, path) => 0;
|
lfs_remove(&lfs, path) => 0;
|
||||||
lfs_file_open(&lfs, &file, path,
|
lfs_file_open(&lfs, &file, path,
|
||||||
@@ -189,7 +189,7 @@ scripts/test.py << TEST
|
|||||||
|
|
||||||
memcpy(path, "coffee/", strlen("coffee/"));
|
memcpy(path, "coffee/", strlen("coffee/"));
|
||||||
memset(path+strlen("coffee/"), 'w', LFS_NAME_MAX);
|
memset(path+strlen("coffee/"), 'w', LFS_NAME_MAX);
|
||||||
path[strlen("coffee/")+LFS_NAME_MAX+1] = '\0';
|
path[strlen("coffee/")+LFS_NAME_MAX] = '\0';
|
||||||
lfs_mkdir(&lfs, path) => 0;
|
lfs_mkdir(&lfs, path) => 0;
|
||||||
lfs_remove(&lfs, path) => 0;
|
lfs_remove(&lfs, path) => 0;
|
||||||
lfs_file_open(&lfs, &file, path,
|
lfs_file_open(&lfs, &file, path,
|
||||||
|
|||||||
@@ -107,6 +107,57 @@ scripts/test.py << TEST
|
|||||||
lfs_unmount(&lfs) => 0;
|
lfs_unmount(&lfs) => 0;
|
||||||
TEST
|
TEST
|
||||||
|
|
||||||
|
echo "--- Write, truncate, and read ---"
|
||||||
|
scripts/test.py << TEST
|
||||||
|
lfs_mount(&lfs, &cfg) => 0;
|
||||||
|
lfs_file_open(&lfs, &file, "sequence",
|
||||||
|
LFS_O_RDWR | LFS_O_CREAT | LFS_O_TRUNC) => 0;
|
||||||
|
|
||||||
|
lfs_size_t size = lfs.cfg->cache_size;
|
||||||
|
lfs_size_t qsize = size / 4;
|
||||||
|
uint8_t *wb = buffer;
|
||||||
|
uint8_t *rb = buffer + size;
|
||||||
|
for (lfs_off_t j = 0; j < size; ++j) {
|
||||||
|
wb[j] = j;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Spread sequence over size */
|
||||||
|
lfs_file_write(&lfs, &file, wb, size) => size;
|
||||||
|
lfs_file_size(&lfs, &file) => size;
|
||||||
|
lfs_file_tell(&lfs, &file) => size;
|
||||||
|
|
||||||
|
lfs_file_seek(&lfs, &file, 0, LFS_SEEK_SET) => 0;
|
||||||
|
lfs_file_tell(&lfs, &file) => 0;
|
||||||
|
|
||||||
|
/* Chop off the last quarter */
|
||||||
|
lfs_size_t trunc = size - qsize;
|
||||||
|
lfs_file_truncate(&lfs, &file, trunc) => 0;
|
||||||
|
lfs_file_tell(&lfs, &file) => 0;
|
||||||
|
lfs_file_size(&lfs, &file) => trunc;
|
||||||
|
|
||||||
|
/* Read should produce first 3/4 */
|
||||||
|
lfs_file_read(&lfs, &file, rb, size) => trunc;
|
||||||
|
memcmp(rb, wb, trunc) => 0;
|
||||||
|
|
||||||
|
/* Move to 1/4 */
|
||||||
|
lfs_file_size(&lfs, &file) => trunc;
|
||||||
|
lfs_file_seek(&lfs, &file, qsize, LFS_SEEK_SET) => qsize;
|
||||||
|
lfs_file_tell(&lfs, &file) => qsize;
|
||||||
|
|
||||||
|
/* Chop to 1/2 */
|
||||||
|
trunc -= qsize;
|
||||||
|
lfs_file_truncate(&lfs, &file, trunc) => 0;
|
||||||
|
lfs_file_tell(&lfs, &file) => qsize;
|
||||||
|
lfs_file_size(&lfs, &file) => trunc;
|
||||||
|
|
||||||
|
/* Read should produce second quarter */
|
||||||
|
lfs_file_read(&lfs, &file, rb, size) => trunc - qsize;
|
||||||
|
memcmp(rb, wb + qsize, trunc - qsize) => 0;
|
||||||
|
|
||||||
|
lfs_file_close(&lfs, &file) => 0;
|
||||||
|
lfs_unmount(&lfs) => 0;
|
||||||
|
TEST
|
||||||
|
|
||||||
echo "--- Truncate and write ---"
|
echo "--- Truncate and write ---"
|
||||||
scripts/test.py << TEST
|
scripts/test.py << TEST
|
||||||
lfs_mount(&lfs, &cfg) => 0;
|
lfs_mount(&lfs, &cfg) => 0;
|
||||||
|
|||||||
Reference in New Issue
Block a user