Mailing List
Home
Forum Home
Maven - Project building tool
Axis - Java SOAP implementation
Lucene - Full-featured text search engine APIs
Cocoon - MVC web framework based on XML/XSL
Fop - Create PDF, PCL, PS, SVG, XML driven by XSL formatting objects.
Log4J - A log library
POI - Java Excel, Word and other Microsoft Office files manipulating library
Oracle database error code ...
Subjects
log4j warning: No appenders could be found
java security AccessControlException: access denied (java io FilePermission clie
java lang InstantiationException: org apache tools ant Main
Apache Axis Tutorial
Subject: Struts <logic iterate >
log4j properties How to parse outpu to multiple files
configuring log4j with BEA Weblogic 8 1
How to use XSL FOP Java together
JSP precompile
Proposal: Adding jar manifest classpath in jar and war plugins
Servlet File Download dialog problem (IE6,Adobe 6 0)
java security AccessControlException: access denied (java io FilePermission
Unsupported major minor version 48 0 problem while running the an
   telope task
Subject: axis wsdl2java Ant Task usage
net sf hibernate MappingException: Error reading resource: test/User hbm xml
Building EAR ANT Script for websphere 5 0
CREATING WAR Files
Classpath problem
jsp data into Excel
Jboss 3 2 3+ vs Tomcat Axis Question
RE: How to include jars and add them into the MANIFEST MF/Class Path
attribute
Printing problem
Subject: InstantiationException
Couldn 't find trusted certificate
Please : How can one install ant 1 6 0 under Eclipse 2 1 ?
Excel: Too many different cell formats
Subject: AXIS: tomcat timeout ?
1 3 final: now giving me java io FileNotFoundException (Too many
open files)
XDoclet, Struts and Maven: Where to start? SOLUTION
Subject: Running junit tests fails
 
Lucene
Page 8 of 247 1   2   3   4   5   6   7   8   9   10   Next 10   Next 100  

Subset match query?

I understand how that recommendation could potentially cover fields with undesired terms mixed in with the desired terms. I fail to see that it covers the case where the undesired term(s) are

Subset match query?

I understand how that recommendation could potentially cover fields with undesired terms mixed in with the desired terms. I fail to see that it covers the case where the undesired term(s) are last i

Subset match query?

A typical solution to problems in this "space " is to index marker terms to denote boundaries in the term sequence ... in combination with things like SpanNear and SpanNot this can be used to make

Subset match query?

Oops too quick to reply... coord() won 't quite do it since it does terms matched in doc versus terms in query. On Oct 3 2007 at 2 20 PM Kyle Maxwell wrote > I 'm indexing a dataset with lot

Subset match query?

See the Similarity.coord() method. /** Computes a score factor based on the fraction of all query terms that a * document contains. This value is multiplied into scores. * * <p >The pre

Subset match query?

I 'm indexing a dataset with lots of short fields. I have determined that it would be useful to highly boost matches where every term in this field is represented in the query. i.e. Query lucene fie

Subject: Re: BoostingTermQuery performance

https //issues.apache.org/jira/browse/LUCENE-1017 On Oct 2 2007 at 8 25 PM Mike Klaas wrote > On 2-Oct-07 at 3 44 PM Peter Keegan wrote > > > I have been experimenting with payloads and Boost

Subject: Re: BoostingTermQuery performance

On 2-Oct-07 at 3 44 PM Peter Keegan wrote > I have been experimenting with payloads and BoostingTermQuery > which I think > are excellent additions to Lucene core. Currently > BoostingTermQu

Subject: Re: BoostingTermQuery performance

Hi Peter This sounds interesting. Can you put this in JIRA as a patch please? I am slowly but surely working on Span query stuff so hopefully I can get to it soon. Thanks Grant On Oct 2 2

Subject: Re: BoostingTermQuery performance

3 okt 2007 kl. 00.44 skrev Peter Keegan > > TermQuery 200 qps > BoostingTermQuery (extends SpanQuery) 97 qps > BoostingTermQuery (extends TermQuery) 130 qps > > Here is a version of BoostingT

Subject: BoostingTermQuery performance

I have been experimenting with payloads and BoostingTermQuery which I think are excellent additions to Lucene core. Currently BoostingTermQuery extends SpanQuery. I would suggest changing this class

Index Dedupe

http //www.gossamer-threads.com/lists/lucene/java-dev/53351 might be of interest. On Oct 1 2007 at 10 25 PM Johnny R. Ruiz III wrote > Hi > > I can 't seem to find a way to delete duplicate in

Index Dedupe

Here 's a couple of fragments alter to suit.... public void doRemove(Directory dir) throws Exception { this.reader IndexReader.open(dir) TermEnum theTerms this.reader.term

Index Dedupe

Hi Daniel Tnx but forgive my ignorance.. can u give me a sample code to do it ). I have never used termDocs() before. Tnx Johnny ----- Original Message ---- From Daniel Noll <daniel@(protected)

Index Dedupe

On Tuesday 02 October 2007 12 25 47 Johnny R. Ruiz III wrote > Hi > > I can 't seem to find a way to delete duplicate in lucene index. I hve a > unique key so it seems to be straight forward. But

Index Dedupe

Hi I can 't seem to find a way to delete duplicate in lucene index. I hve a unique key so it seems to be straight forward. But I can 't find a simple way to do it except for putting each record i

mixing analyzer

The whole question of multilingual indexing has been discusses at length you might find some ideas if you search the archive... Erick On 10/1/07 Dino Korah <dckorah@(protected) > wrote > > Thanks Er

Subject: Re: Indexing puncuation and symbols

You might be able to create an analyzer that breaks your stream up (from the example) into tokens "foo " and " " and then (using the same analyzer) search on phrases with a slop of 0. That seems like i

Subject: GOMStaxWriter compile error

I 've been getting the following compiler error when building the javadocs from the trunk sources Ant build error [javac] D \lucene- 2.2.0\contrib\gdata-server\src\gom\src\java\org\apache\lucen

mixing analyzer

Thanks Erick. The PerFieldAnalyzerWrapper could fit in but in the current world of multilingual anywhere (even in programming languages.. %$?%#@) almost any field in an email (addresses subject b

Subject: Re: a query for a special AND?

As for suggestions on how to do this I have no other than to make sure that you can create the queries necessary to obtain the required output. Regards Paul Elschot On Sunday 30 September 2007 09

Subject: Re: Indexing puncuation and symbols

Of course it depends on the kind of query you are doing but (I did find the query parser in the mean time) MultiFieldQueryParser mfqp new MultiFieldQueryParser(useFields analyzer boosts) where

Subject: Re: Indexing puncuation and symbols

Well the size wouldn 't be a problem we could afford the extra field. But it would seem to complicate the search quite a lot. I 'd have to run the search terms through both analyzers. It would be mu

Subject: Re: Indexing puncuation and symbols

Hi Don 't know the size of your dataset. But couldn 't you index in 2 fields with PerFieldAnalyzer tokenizing with Standard for 1 field and WhiteSpace for the other. Then use multiple field query

Subject: Re: Indexing puncuation and symbols

Whitespace analyzer does preserve those symbols but not as tokens. It simply leaves them attached to the original term. As an example of what I 'm talking about consider a document that contains (

Subject: Re: Indexing puncuation and symbols

1 okt 2007 kl. 15.33 skrev John Byrne > Has anyone written an analyzer that preserves puncuation and > synmbols ( "? " "$ " "% " etc.) as tokens? WhitespaceAnalyzer? You could also extend the lexic

Some questions...

1 okt 2007 kl. 14.41 skrev sandeep chawla > 2- Is there a way I can get the term.docFrq() for a particular set of > documents.. Using TermDocs or the TermFreqVector. -- karl

Subject: Indexing puncuation and symbols

Hi Has anyone written an analyzer that preserves puncuation and synmbols ( "? " "$ " "% " etc.) as tokens? That way we could distinguish between searching for "100 " and "100% " or "$100 ". Does anyo

mixing analyzer

Sure but there 's a time/space tradeoff. Isn 't there always <G >.... PerFieldAnalyzerWrapper is your friend. It would require that your index be built on a per-language basis. Say indexing text from F

mixing analyzer

Hi I am working on a lucene email indexing system which potentially can get documents in various languages. Currently I am using StandardAnalyzer which works for English but not for many of the
Page 8 of 247 1   2   3   4   5   6   7   8   9   10   Next 10   Next 100