-
Notifications
You must be signed in to change notification settings - Fork 17
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
hyp drive sync . -y
crashes with heap out of memory allocation failure
#53
Comments
I was able to use the hyperdrive api directly to import this archive, so I'm pretty sure this memory issue is in hyp, one of the layers on top of hyperdrive that hyp uses, or in how the traversal for files happens. This script runs to import the archive with very little memory use: #!/usr/bin/env node
var Hyperdrive = require('hyperdrive')
var fs = require('fs')
var path = require('path')
var pump = require('pumpify')
var minimist = require('minimist')
var argv = minimist(process.argv.slice(2))
if (argv._[0] === 'import') {
var hdir = argv.o || '.hyperdrive'
var drive = new Hyperdrive(hdir)
var cursors = [argv._[1]]
var count = { files: 0, directories: 1 }
var errors = []
var prevLines = 0
var iv = setInterval(print, 1000)
function error(err) {
errors.push({ time: Date.now(), error: err })
}
function print() {
var clear = '\x1b[K'
for (var i = 0; i < prevLines; i++) {
clear += '\x1b[1A\x1b[K'
}
prevLines = 1 + errors.length
var nFile = fmt(count.files).padStart(9)
var nDir = fmt(count.directories).padStart(9)
console.log(`${clear}${nFile} files, ${nDir} directories`)
errors.forEach(function (e) {
console.log('Error: ' + e.error.message)
})
errors = errors.filter(e => Date.now() - e.time < 5*60_000) // show for 5 minutes
}
;(function next() {
if (cursors.length === 0) {
clearInterval(iv)
print()
return
}
var c = cursors.shift()
fs.readdir(c, function (err, files) {
if (err) {
error(err)
return next()
}
files = files.filter(file => !/^\./.test(file))
var pending = 1 + files.length
files.forEach(function (f) {
var file = path.join(c,f)
fs.stat(file, function (err, stat) {
if (err) {
error(err)
if (--pending === 0) next()
return
} else if (stat.isDirectory()) {
drive.mkdir(file, function (err) {
if (err) error(err)
else {
count.directories++
cursors.push(file)
}
if (--pending === 0) next()
})
} else {
fs.readFile(file, function (err, src) {
if (err) {
error(err)
if (--pending === 0) next()
return
}
drive.writeFile(file, src, function (err) {
if (err) error(err)
else count.files++
if (--pending === 0) next()
})
})
}
})
})
if (--pending === 0) next()
})
})()
} else if (argv._[0] === 'addr') {
var drive = new Hyperdrive(argv.o || argv._[1] || '.hyperdrive')
drive.once('ready', function () {
console.log(`hyper://${drive.key.toString('hex')}`)
})
} else if (argv._[0] === 'share') {
var drive = new Hyperdrive(argv.o || argv._[1] || '.hyperdrive')
drive.once('ready', function () {
console.log(`hyper://${drive.key.toString('hex')}`)
var swarm = require('hyperswarm')()
var crypto = require('crypto')
swarm.join(drive.discoveryKey, { lookup: true, announce: true })
swarm.on('connection', (socket, info) => {
socket.on('error', err => console.error(err))
pump(socket, drive.replicate(info.client), socket)
})
swarm.on('error', err => console.error(err))
})
}
function fmt(x) {
return x.toString()
.split('')
.reverse()
.join('')
.replace(/(\d{3})/g,'$1_')
.split('')
.reverse()
.join('')
.replace(/^_/,'')
} |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Labels
None yet
0 participants
When I try to seed a directory with 99GB of data spread across 2.6 million files, I get a memory allocation error after 2 minutes:
Watching
htop
the whole time, the highest I saw the memory usage go was up to 10% on this machine with 60GB of memory.To reproduce, the data is also seeded to /ipfs/QmVCYUK51Miz4jEjJxCq3bA6dfq5FXD6s2EYp6LjHQhGmh
The text was updated successfully, but these errors were encountered: