MySQL importing CSV files really slow
I am trying to import CSV files into a table, but it has been very slow. I have about 1000 files with a file size of 40 MB each. Whenever I try to import it, I can see with for example MySQL workbench that it is inserting in a rate of about 30 - 60 rows per second. It will take ages before al my files are processed. How can I speed this up? I have already modified the mysql.cnf file to the following config, which unfortunately does not speed it up:
/etc/mysql/conf.d/mysql.cnf
[mysqld]
innodb_buffer_pool_size=12G
innodb_io_capacity = 2000
innodb_read_io_threads = 48
innodb_thread_concurrency = 0
innodb_write_io_threads = 48
innodb_buffer_pool_size=12G
innodb_log_file_size = 512M
max_connections = 1000
max_allowed_packet = 128M
#key_buffer = 1000M
bulk_insert_buffer_size = 1024M
innodb_doublewrite = 0
innodb_flush_log_at_trx_commit = 2
key_buffer_size = 4000M
It does say that 12G is configured in InnoDB according to PHPMyAdmin, so I can safely confirm that the config at least works.
Specs in case needed:
OS: Ubuntu 18.04
CPU: 6 cores, 12 threads @ 4.5 GHz
RAM: 32 GB DDR4 @ 3.2 GHz
SSD: 1 TB NVME @ 3.5 GB/s Read & 3.3 GB/s Write
18.04 mysql phpmyadmin mysql-workbench csv
New contributor
add a comment |
I am trying to import CSV files into a table, but it has been very slow. I have about 1000 files with a file size of 40 MB each. Whenever I try to import it, I can see with for example MySQL workbench that it is inserting in a rate of about 30 - 60 rows per second. It will take ages before al my files are processed. How can I speed this up? I have already modified the mysql.cnf file to the following config, which unfortunately does not speed it up:
/etc/mysql/conf.d/mysql.cnf
[mysqld]
innodb_buffer_pool_size=12G
innodb_io_capacity = 2000
innodb_read_io_threads = 48
innodb_thread_concurrency = 0
innodb_write_io_threads = 48
innodb_buffer_pool_size=12G
innodb_log_file_size = 512M
max_connections = 1000
max_allowed_packet = 128M
#key_buffer = 1000M
bulk_insert_buffer_size = 1024M
innodb_doublewrite = 0
innodb_flush_log_at_trx_commit = 2
key_buffer_size = 4000M
It does say that 12G is configured in InnoDB according to PHPMyAdmin, so I can safely confirm that the config at least works.
Specs in case needed:
OS: Ubuntu 18.04
CPU: 6 cores, 12 threads @ 4.5 GHz
RAM: 32 GB DDR4 @ 3.2 GHz
SSD: 1 TB NVME @ 3.5 GB/s Read & 3.3 GB/s Write
18.04 mysql phpmyadmin mysql-workbench csv
New contributor
How do you do the inserts? One by one? Does the table have any indexes?
– vidarlo
9 hours ago
The inserts go one by one (Using Jetbrains Datagrip for it as phpmyadmin is very slow), and the table does not have any index to my knowledge
– Mark D
9 hours ago
add a comment |
I am trying to import CSV files into a table, but it has been very slow. I have about 1000 files with a file size of 40 MB each. Whenever I try to import it, I can see with for example MySQL workbench that it is inserting in a rate of about 30 - 60 rows per second. It will take ages before al my files are processed. How can I speed this up? I have already modified the mysql.cnf file to the following config, which unfortunately does not speed it up:
/etc/mysql/conf.d/mysql.cnf
[mysqld]
innodb_buffer_pool_size=12G
innodb_io_capacity = 2000
innodb_read_io_threads = 48
innodb_thread_concurrency = 0
innodb_write_io_threads = 48
innodb_buffer_pool_size=12G
innodb_log_file_size = 512M
max_connections = 1000
max_allowed_packet = 128M
#key_buffer = 1000M
bulk_insert_buffer_size = 1024M
innodb_doublewrite = 0
innodb_flush_log_at_trx_commit = 2
key_buffer_size = 4000M
It does say that 12G is configured in InnoDB according to PHPMyAdmin, so I can safely confirm that the config at least works.
Specs in case needed:
OS: Ubuntu 18.04
CPU: 6 cores, 12 threads @ 4.5 GHz
RAM: 32 GB DDR4 @ 3.2 GHz
SSD: 1 TB NVME @ 3.5 GB/s Read & 3.3 GB/s Write
18.04 mysql phpmyadmin mysql-workbench csv
New contributor
I am trying to import CSV files into a table, but it has been very slow. I have about 1000 files with a file size of 40 MB each. Whenever I try to import it, I can see with for example MySQL workbench that it is inserting in a rate of about 30 - 60 rows per second. It will take ages before al my files are processed. How can I speed this up? I have already modified the mysql.cnf file to the following config, which unfortunately does not speed it up:
/etc/mysql/conf.d/mysql.cnf
[mysqld]
innodb_buffer_pool_size=12G
innodb_io_capacity = 2000
innodb_read_io_threads = 48
innodb_thread_concurrency = 0
innodb_write_io_threads = 48
innodb_buffer_pool_size=12G
innodb_log_file_size = 512M
max_connections = 1000
max_allowed_packet = 128M
#key_buffer = 1000M
bulk_insert_buffer_size = 1024M
innodb_doublewrite = 0
innodb_flush_log_at_trx_commit = 2
key_buffer_size = 4000M
It does say that 12G is configured in InnoDB according to PHPMyAdmin, so I can safely confirm that the config at least works.
Specs in case needed:
OS: Ubuntu 18.04
CPU: 6 cores, 12 threads @ 4.5 GHz
RAM: 32 GB DDR4 @ 3.2 GHz
SSD: 1 TB NVME @ 3.5 GB/s Read & 3.3 GB/s Write
18.04 mysql phpmyadmin mysql-workbench csv
18.04 mysql phpmyadmin mysql-workbench csv
New contributor
New contributor
New contributor
asked 10 hours ago
Mark DMark D
1314
1314
New contributor
New contributor
How do you do the inserts? One by one? Does the table have any indexes?
– vidarlo
9 hours ago
The inserts go one by one (Using Jetbrains Datagrip for it as phpmyadmin is very slow), and the table does not have any index to my knowledge
– Mark D
9 hours ago
add a comment |
How do you do the inserts? One by one? Does the table have any indexes?
– vidarlo
9 hours ago
The inserts go one by one (Using Jetbrains Datagrip for it as phpmyadmin is very slow), and the table does not have any index to my knowledge
– Mark D
9 hours ago
How do you do the inserts? One by one? Does the table have any indexes?
– vidarlo
9 hours ago
How do you do the inserts? One by one? Does the table have any indexes?
– vidarlo
9 hours ago
The inserts go one by one (Using Jetbrains Datagrip for it as phpmyadmin is very slow), and the table does not have any index to my knowledge
– Mark D
9 hours ago
The inserts go one by one (Using Jetbrains Datagrip for it as phpmyadmin is very slow), and the table does not have any index to my knowledge
– Mark D
9 hours ago
add a comment |
2 Answers
2
active
oldest
votes
You should have a look at the LOAD DATA
function of MySQL, or the mysqlimport
tool.
Those two functions are reportedly 20-30 times faster than insert
'ing one by one, which would bring you up to at least 500-1000 rows a second.
As you don't provide any information about your data format, it's hard to give a detailed use instruction for mysqlimport
, which seems to be the most applicable tool in your situation.
add a comment |
It turned out that the file encoding was different between each file. Some files had UTF-8 encoding, some had ISO 8859-1 (Latin1) encoding, which caused a lot of programs to take ages to import due to (I think) converting it into another encoding. A 40 MB file now only takes about 6 seconds to import instead of 48 minutes.. Thanks @vidarlo for the suggestion of mysqlimport, which is certainly an improvement, but overall the time improvement is good enough for me!
New contributor
add a comment |
Your Answer
StackExchange.ready(function() {
var channelOptions = {
tags: "".split(" "),
id: "89"
};
initTagRenderer("".split(" "), "".split(" "), channelOptions);
StackExchange.using("externalEditor", function() {
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled) {
StackExchange.using("snippets", function() {
createEditor();
});
}
else {
createEditor();
}
});
function createEditor() {
StackExchange.prepareEditor({
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: true,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
imageUploader: {
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
},
onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
});
}
});
Mark D is a new contributor. Be nice, and check out our Code of Conduct.
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2faskubuntu.com%2fquestions%2f1124589%2fmysql-importing-csv-files-really-slow%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
2 Answers
2
active
oldest
votes
2 Answers
2
active
oldest
votes
active
oldest
votes
active
oldest
votes
You should have a look at the LOAD DATA
function of MySQL, or the mysqlimport
tool.
Those two functions are reportedly 20-30 times faster than insert
'ing one by one, which would bring you up to at least 500-1000 rows a second.
As you don't provide any information about your data format, it's hard to give a detailed use instruction for mysqlimport
, which seems to be the most applicable tool in your situation.
add a comment |
You should have a look at the LOAD DATA
function of MySQL, or the mysqlimport
tool.
Those two functions are reportedly 20-30 times faster than insert
'ing one by one, which would bring you up to at least 500-1000 rows a second.
As you don't provide any information about your data format, it's hard to give a detailed use instruction for mysqlimport
, which seems to be the most applicable tool in your situation.
add a comment |
You should have a look at the LOAD DATA
function of MySQL, or the mysqlimport
tool.
Those two functions are reportedly 20-30 times faster than insert
'ing one by one, which would bring you up to at least 500-1000 rows a second.
As you don't provide any information about your data format, it's hard to give a detailed use instruction for mysqlimport
, which seems to be the most applicable tool in your situation.
You should have a look at the LOAD DATA
function of MySQL, or the mysqlimport
tool.
Those two functions are reportedly 20-30 times faster than insert
'ing one by one, which would bring you up to at least 500-1000 rows a second.
As you don't provide any information about your data format, it's hard to give a detailed use instruction for mysqlimport
, which seems to be the most applicable tool in your situation.
answered 9 hours ago
vidarlovidarlo
10.9k52750
10.9k52750
add a comment |
add a comment |
It turned out that the file encoding was different between each file. Some files had UTF-8 encoding, some had ISO 8859-1 (Latin1) encoding, which caused a lot of programs to take ages to import due to (I think) converting it into another encoding. A 40 MB file now only takes about 6 seconds to import instead of 48 minutes.. Thanks @vidarlo for the suggestion of mysqlimport, which is certainly an improvement, but overall the time improvement is good enough for me!
New contributor
add a comment |
It turned out that the file encoding was different between each file. Some files had UTF-8 encoding, some had ISO 8859-1 (Latin1) encoding, which caused a lot of programs to take ages to import due to (I think) converting it into another encoding. A 40 MB file now only takes about 6 seconds to import instead of 48 minutes.. Thanks @vidarlo for the suggestion of mysqlimport, which is certainly an improvement, but overall the time improvement is good enough for me!
New contributor
add a comment |
It turned out that the file encoding was different between each file. Some files had UTF-8 encoding, some had ISO 8859-1 (Latin1) encoding, which caused a lot of programs to take ages to import due to (I think) converting it into another encoding. A 40 MB file now only takes about 6 seconds to import instead of 48 minutes.. Thanks @vidarlo for the suggestion of mysqlimport, which is certainly an improvement, but overall the time improvement is good enough for me!
New contributor
It turned out that the file encoding was different between each file. Some files had UTF-8 encoding, some had ISO 8859-1 (Latin1) encoding, which caused a lot of programs to take ages to import due to (I think) converting it into another encoding. A 40 MB file now only takes about 6 seconds to import instead of 48 minutes.. Thanks @vidarlo for the suggestion of mysqlimport, which is certainly an improvement, but overall the time improvement is good enough for me!
New contributor
New contributor
answered 8 hours ago
Mark DMark D
1314
1314
New contributor
New contributor
add a comment |
add a comment |
Mark D is a new contributor. Be nice, and check out our Code of Conduct.
Mark D is a new contributor. Be nice, and check out our Code of Conduct.
Mark D is a new contributor. Be nice, and check out our Code of Conduct.
Mark D is a new contributor. Be nice, and check out our Code of Conduct.
Thanks for contributing an answer to Ask Ubuntu!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2faskubuntu.com%2fquestions%2f1124589%2fmysql-importing-csv-files-really-slow%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
How do you do the inserts? One by one? Does the table have any indexes?
– vidarlo
9 hours ago
The inserts go one by one (Using Jetbrains Datagrip for it as phpmyadmin is very slow), and the table does not have any index to my knowledge
– Mark D
9 hours ago