[Groonga-commit] groonga/groonga at 7a28e81 [master] tokenize: remove table option

Zurück zum Archiv-Index

naoa null+****@clear*****
Sun Oct 26 18:27:34 JST 2014


naoa	2014-10-26 18:27:34 +0900 (Sun, 26 Oct 2014)

  New Revision: 7a28e8167b47c7f3f1fc6325c2bdf44873a52154
  https://github.com/groonga/groonga/commit/7a28e8167b47c7f3f1fc6325c2bdf44873a52154

  Merged ff11b0c: Merge pull request #232 from naoa/tokenize-support-token_filters

  Message:
    tokenize: remove table option

  Modified files:
    lib/proc.c
    test/command/suite/tokenize/with_token_filters.expected
    test/command/suite/tokenize/with_token_filters.test

  Modified: lib/proc.c (+5 -12)
===================================================================
--- lib/proc.c    2014-10-26 17:24:38 +0900 (15948b6)
+++ lib/proc.c    2014-10-26 18:27:34 +0900 (74b85d6)
@@ -3545,7 +3545,6 @@ proc_tokenize(grn_ctx *ctx, int nargs, grn_obj **args, grn_user_data *user_data)
   grn_obj *flag_names;
   grn_obj *mode_name;
   grn_obj *token_filter_names;
-  grn_obj *table_name;
 
   tokenizer_name = VAR(0);
   string = VAR(1);
@@ -3553,7 +3552,6 @@ proc_tokenize(grn_ctx *ctx, int nargs, grn_obj **args, grn_user_data *user_data)
   flag_names = VAR(3);
   mode_name = VAR(4);
   token_filter_names = VAR(5);
-  table_name = VAR(6);
 
   if (GRN_TEXT_LEN(tokenizer_name) == 0) {
     ERR(GRN_INVALID_ARGUMENT, "[tokenize] tokenizer name is missing");
@@ -3574,14 +3572,10 @@ proc_tokenize(grn_ctx *ctx, int nargs, grn_obj **args, grn_user_data *user_data)
       return NULL;
     }
 
-    if (GRN_TEXT_LEN(table_name)) {
-      lexicon = grn_ctx_get(ctx, GRN_TEXT_VALUE(table_name), GRN_TEXT_LEN(table_name));
-    } else {
-      lexicon = create_lexicon_for_tokenize(ctx,
-                                            tokenizer_name,
-                                            normalizer_name,
-                                            token_filter_names);
-    }
+    lexicon = create_lexicon_for_tokenize(ctx,
+                                          tokenizer_name,
+                                          normalizer_name,
+                                          token_filter_names);
     if (!lexicon) {
       return NULL;
     }
@@ -5499,8 +5493,7 @@ grn_db_init_builtin_query(grn_ctx *ctx)
   DEF_VAR(vars[3], "flags");
   DEF_VAR(vars[4], "mode");
   DEF_VAR(vars[5], "token_filters");
-  DEF_VAR(vars[6], "table");
-  DEF_COMMAND("tokenize", proc_tokenize, 7, vars);
+  DEF_COMMAND("tokenize", proc_tokenize, 6, vars);
 
   DEF_COMMAND("tokenizer_list", proc_tokenizer_list, 0, vars);
 

  Modified: test/command/suite/tokenize/with_token_filters.expected (+7 -20)
===================================================================
--- test/command/suite/tokenize/with_token_filters.expected    2014-10-26 17:24:38 +0900 (87c7367)
+++ test/command/suite/tokenize/with_token_filters.expected    2014-10-26 18:27:34 +0900 (b5ce78f)
@@ -1,15 +1,6 @@
-register token_filters/stop_word
+register token_filters/stem
 [[0,0.0,0.0],true]
-table_create Terms TABLE_PAT_KEY ShortText   --default_tokenizer TokenBigram   --normalizer NormalizerAuto   --token_filters TokenFilterStopWord
-[[0,0.0,0.0],true]
-column_create Terms is_stop_word COLUMN_SCALAR Bool
-[[0,0.0,0.0],true]
-load --table Terms
-[
-{"_key": "and", "is_stop_word": true}
-]
-[[0,0.0,0.0],1]
-tokenize TokenBigram "Hello and Good-bye" --mode GET --table Terms
+tokenize TokenBigram "I developed Groonga" NormalizerAuto --token_filters TokenFilterStem
 [
   [
     0,
@@ -18,20 +9,16 @@ tokenize TokenBigram "Hello and Good-bye" --mode GET --table Terms
   ],
   [
     {
-      "value": "hello",
+      "value": "i",
       "position": 0
     },
     {
-      "value": "good",
-      "position": 2
-    },
-    {
-      "value": "-",
-      "position": 3
+      "value": "develop",
+      "position": 1
     },
     {
-      "value": "bye",
-      "position": 4
+      "value": "groonga",
+      "position": 2
     }
   ]
 ]

  Modified: test/command/suite/tokenize/with_token_filters.test (+4 -13)
===================================================================
--- test/command/suite/tokenize/with_token_filters.test    2014-10-26 17:24:38 +0900 (19a85a5)
+++ test/command/suite/tokenize/with_token_filters.test    2014-10-26 18:27:34 +0900 (899e966)
@@ -1,14 +1,5 @@
-register token_filters/stop_word
+#@on-error omit
+register token_filters/stem
+#@on-error default
 
-table_create Terms TABLE_PAT_KEY ShortText \
-  --default_tokenizer TokenBigram \
-  --normalizer NormalizerAuto \
-  --token_filters TokenFilterStopWord
-column_create Terms is_stop_word COLUMN_SCALAR Bool
-
-load --table Terms
-[
-{"_key": "and", "is_stop_word": true}
-]
-
-tokenize TokenBigram "Hello and Good-bye" --mode GET --table Terms
+tokenize TokenBigram "I developed Groonga" NormalizerAuto --token_filters TokenFilterStem
-------------- next part --------------
HTML����������������������������...
Download 



More information about the Groonga-commit mailing list
Zurück zum Archiv-Index