473,387 Members | 1,721 Online
Bytes | Software Development & Data Engineering Community
Post Job

Home Posts Topics Members FAQ

Join Bytes to post your question to a community of 473,387 software developers and data experts.

Best text search for many Gigabyte Database?

I have a so big DB, It's about 10-100GB of text.
Now I want to create a searching function but It take too long when I
use LIKE from MySQL
Anyone know the best sollution?
Can I use Google Desktop search or an others?
Please help!

Feb 5 '07 #1
5 2816
ngocviet wrote:
I have a so big DB, It's about 10-100GB of text.
Now I want to create a searching function but It take too long when I
use LIKE from MySQL
Anyone know the best sollution?
Can I use Google Desktop search or an others?
Please help!
If you use mySQL, consider using FULLTEXT and MATCH AGAINST().
Bottomline stays: If you have 100GB of data, some searchtime is needed, but
using FULLTEXT can speed that up considerable.
Check your mySQL documentation for details.

Regards,
Erwin Moller
Feb 5 '07 #2
If you use mySQL, consider using FULLTEXT and MATCH AGAINST().
Bottomline stays: If you have 100GB of data, some searchtime is needed, but
using FULLTEXT can speed that up considerable.
Check your mySQL documentation for details.
I've tried FULLTEXT but it take too long, about over 10 seconds with
1.2 Gigabyte table.
If need, I can convert database to other structure.

Feb 5 '07 #3
ngocviet wrote:
>
>If you use mySQL, consider using FULLTEXT and MATCH AGAINST().
Bottomline stays: If you have 100GB of data, some searchtime is needed,
but using FULLTEXT can speed that up considerable.
Check your mySQL documentation for details.

I've tried FULLTEXT but it take too long, about over 10 seconds with
1.2 Gigabyte table.
If need, I can convert database to other structure.
Hi,

If you use MySQL ISAM with FULLTEXT, and not LIKE, but MATCH AGAINST, you
are using one of the fastest approaches a developer can set up (as far as I
know).
I expect the only way to increase searchspeed for a 100GB database filled
with text, is throwing more/better hardware at it (more memory, faster disk
IO, faster CPU, etc.).
In general: When searching through a huge datastructure: Disk IO is the
bottleneck. So faster disk IO will help the most.
Dive into different RAID systems maybe.
eg: If you have 2 HD delivering data at the same time (using some RAID),
your query will maybe run twice as fast.

Regards,
Erwin Moller
Feb 5 '07 #4
"ngocviet" <ng******@gmail.comwrote in message
news:11*********************@h3g2000cwc.googlegrou ps.com...
>
>If you use mySQL, consider using FULLTEXT and MATCH AGAINST().
Bottomline stays: If you have 100GB of data, some searchtime is needed,
but
using FULLTEXT can speed that up considerable.
Check your mySQL documentation for details.

I've tried FULLTEXT but it take too long, about over 10 seconds with
1.2 Gigabyte table.
If need, I can convert database to other structure.

A search like that for such an amount of unindexed textual data simply has
to take some time. Live with it, or invest to hardware. Like the IT proverb
says: "If it doesn't work, just throw money at it".

--
"Ohjelmoija on organismi joka muuttaa kofeiinia koodiksi" - lpk
http://outolempi.net/ahdistus/ - Satunnaisesti päivittyvä nettisarjis
sp**@outolempi.net | rot13(xv***@bhgbyrzcv.arg)
Feb 5 '07 #5
In addition to MySQL FULLTEXT, you could also use Lucene
(Zend_Search_Lucene in PHP or perhaps Solr running inside Tomcat).
Performance is quite good and Lucene is very flexible.

On Feb 5, 1:27 am, "ngocviet" <ngocv...@gmail.comwrote:
I have a so big DB, It's about 10-100GB of text.
Now I want to create a searching function but It take too long when I
use LIKE from MySQL
Anyone know the best sollution?
Can I use Google Desktop search or an others?
Please help!

Feb 6 '07 #6

This thread has been closed and replies have been disabled. Please start a new discussion.

Similar topics

10
by: Bob Bedford | last post by:
Sorry if I post here, but I don't have access to any valuable mysql ng for this question, ans since 99% or php programmers deal with mysql, I'm sure I'll get an answer here. Look at the...
10
by: Lorn Davies | last post by:
Hi there, I'm a Python newbie hoping for some direction in working with text files that range from 100MB to 1G in size. Basically certain rows, sorted by the first (primary) field maybe second...
60
by: Julie | last post by:
What is the *fastest* way in .NET to search large on-disk text files (100+ MB) for a given string. The files are unindexed and unsorted, and for the purposes of my immediate requirements, can't...
0
by: Anonieko Ramos | last post by:
ASP.NET Forms Authentication Best Practices Dr. Dobb's Journal February 2004 Protecting user information is critical By Douglas Reilly Douglas is the author of Designing Microsoft ASP.NET...
8
by: robert | last post by:
Hello, I want to put (incrementally) changed/new files from a big file tree "directly,compressed and password-only-encrypted" to a remote backup server incrementally via FTP,SFTP or DAV.... At...
1
by: jrs_14618 | last post by:
Hello All, This post is essentially a reply a previous post/thread here on this mailing.database.myodbc group titled: MySQL 4.0, FULL-TEXT Indexing and Search Arabic Data, Unicode I was...
5
by: Rich | last post by:
Hello, I have a search application to search data in tables in a database (3 sql server tables). I populate 2 comboboxes with with data from each table. One combobox will contain unique...
3
by: Russell | last post by:
Hey, ok i have numerous tables to search through for a 'site search'. some of the searchble fields have html embeded within so after some quick referencing, saw I can use the regExp function...
13
by: G | last post by:
Hello, Looking for opinions on a fairly simple task, new to ASP.net (C#) and want to make sure I do this as efficiently as possible. I have a web based form, and I need to run some SQL before...
0
by: ryjfgjl | last post by:
In our work, we often receive Excel tables with data in the same format. If we want to analyze these data, it can be difficult to analyze them because the data is spread across multiple Excel files...
0
by: emmanuelkatto | last post by:
Hi All, I am Emmanuel katto from Uganda. I want to ask what challenges you've faced while migrating a website to cloud. Please let me know. Thanks! Emmanuel
1
by: nemocccc | last post by:
hello, everyone, I want to develop a software for my android phone for daily needs, any suggestions?
1
by: Sonnysonu | last post by:
This is the data of csv file 1 2 3 1 2 3 1 2 3 1 2 3 2 3 2 3 3 the lengths should be different i have to store the data by column-wise with in the specific length. suppose the i have to...
0
by: Hystou | last post by:
There are some requirements for setting up RAID: 1. The motherboard and BIOS support RAID configuration. 2. The motherboard has 2 or more available SATA protocol SSD/HDD slots (including MSATA, M.2...
0
marktang
by: marktang | last post by:
ONU (Optical Network Unit) is one of the key components for providing high-speed Internet services. Its primary function is to act as an endpoint device located at the user's premises. However,...
0
by: Hystou | last post by:
Most computers default to English, but sometimes we require a different language, especially when relocating. Forgot to request a specific language before your computer shipped? No problem! You can...
0
Oralloy
by: Oralloy | last post by:
Hello folks, I am unable to find appropriate documentation on the type promotion of bit-fields when using the generalised comparison operator "<=>". The problem is that using the GNU compilers,...
0
jinu1996
by: jinu1996 | last post by:
In today's digital age, having a compelling online presence is paramount for businesses aiming to thrive in a competitive landscape. At the heart of this digital strategy lies an intricately woven...

By using Bytes.com and it's services, you agree to our Privacy Policy and Terms of Use.

To disable or enable advertisements and analytics tracking please visit the manage ads & tracking page.