Module:User:Suzukaze-c/02
Appearance
- This module sandbox lacks a documentation subpage. You may create it.
- Useful links: root page • root page’s subpages • links • transclusions • testcases • user page • user talk page • userspace
This is a private module sandbox of Suzukaze-c, for their own experimentation. Items in this module may be added and removed at Suzukaze-c's discretion; do not rely on this module's stability.
--[==[
mostly unrelated things, in one single module bc i can.
show
format simple word lists
rpre
convert to {{ja-r}} using data given
extract_ja_readings
extract readings from a japanese entry
r
convert to {{ja-r}} using data in the linked entry
hzreport
generate hanzi entry report
newja
new japanese entry generator
newhz
new hanzi entry generator
newhzmul
new hanzi translingual entry generator, from a copy-paste of the unihan page. how horrifying
test_2
create a special:search link
test_3
newhz() wu
test_9
ltc/och attention
note to self: consider retracting {{zh-attn-split}}
test_10
determine ?action=edit§ion=x
test_11
HSK list words missing appropriate category → https://en.wiktionary.org/?oldid=46157868
test_12
jouyou kanji that need {{ja-readings}} to be updated
[[User:Suzukaze-c/cleanup/12]]
test_13
process/update {{ja-readings}}
test_14
generate list of {{attention}}s from a copy-paste of the Category: page
make_map
map_test_data
kartographer test
test_16
kun readings that may be a verb with a '-' in the wrong place
https://en.wiktionary.org/w/index.php?title=%E6%89%93&diff=prev&oldid=50292857
test_17
look for missing japanese entries, based on jouyou kanji data
https://en.wiktionary.org/?oldid=52318414
test_18
extract entry names from a copy-paste of the Category: page
test_19
do things from a copy-paste of [[Category:Chinese terms with uncreated forms]]
=p.newja{['args']={['testing_pagename']='プログレッシブロック', 'プログレッシブ ロック', 'n'}}
]==]
local export = {}
local replace = mw.ustring.gsub
local find = mw.ustring.find
local match = mw.ustring.match
local itermatch = mw.ustring.gmatch
local split = mw.text.split
local itersplit = mw.text.gsplit
local trim = mw.text.trim
local lower = mw.ustring.lower
local sub = mw.ustring.sub
local len = mw.ustring.len
function export.show(frame)
local text = trim(frame.args[2])
local lang = frame.args[1]
local out = {}
local non = require('Module:string utilities').pattern_escape('*!?.,。、:;…《》「」【】()!?.,:;()"—·☆[] ')
local rubied = false
text = replace(text, '\n+', '*')
text = replace(text, ' ', ' ')
text = replace(text, '%s+', ' ')
if not find(text, '[ *]') then
for char in itermatch(text, '(.)') do
table.insert(out, '[['..char..']] ')
end
else
for punca, word, puncb in itermatch(text, '(['..non..']*)([^'..non..']+)(['..non..']*)') do
if (lang == 'ja') and match(word, '[¥・]') then
rubied = true
word = replace(word, '¥', ' ')
local lemma, kana = match(word, '(.+)・(.+)')
if not kana then
lemma = word
kana = replace(word, '([㐀-鿕々])', '!')
end
word = require('module:ja-link').link({lemma = lemma, kana = kana})
table.insert(out, punca..word..puncb)
else
table.insert(out, punca..'[['..word..']]'..puncb)
end
end
end
out = table.concat(out)
if rubied then
out = require('module:links').full_link({term = out .. '//', tr = '-', lang = require('module:languages').getByCode(lang), sc = require('module:scripts').getByCode('Jpan')})
else
out = require('module:links').full_link({term = out .. '//', tr = '-', lang = require('module:languages').getByCode(lang)})
end
return '«' .. lang .. '» ' .. out
end
function export.rpre(frame)
-- {{m|1=ja|2=WORD|3=TITLE|4=GLOSS}}
-- {{m|0=ja|1=WORD|2=TITLE|3=GLOSS}}
local one = frame.args[1] or ''
local two = frame.args[2] or ''
local three = frame.args[3] or ''
local four = frame.args[4] or ''
local jp = ''
local tr = frame.args['tr'] or ''
local gloss = frame.args['gloss'] or ''
if one == 'ja' then
jp = two
linktitle = three
gloss = (gloss ~= '' and gloss or four)
else
jp = one
linktitle = two
gloss = three
end
tr = replace(tr, '[^¥.^、ぁ-ー]+', '')
tr = replace(tr, '¥', ' ')
if gloss ~= '' then
gloss = ': ' .. gloss
end
if tr ~= '' then
tr = '|' .. tr
end
if tr == '' and find(jp, '[㐀-鿕𠀀-𬺡]') then
tr = '|' .. jp
end
if linktitle ~= '' then
jp = 'linkto=' .. jp .. '|' .. linktitle
end
return '{{ja-r|' .. jp .. tr .. '}}' .. gloss
end
function export.extract_ja_readings(pagename)
if match(pagename, "%[%[") then
--error("Cannot process Japanese text with embedded wikilinks.")
return {}
end
local readings
local content
local function process(text)
text = replace(text, 'hhira=[^|}]+', '')
text = replace(text, 'decl=[^|}]+', '')
text = replace(text, 'infl=[^|}]+', '')
text = replace(text, 'kyu=[^|}]+', '')
text = replace(text, 'head=[^|}]+', '')
text = replace(text, 'hira=', '')
if text == 'proper' or text == 'proper noun' then
table.insert(readings, '^' .. pagename)
end
if find(text, 'proper') and not find(text, '%^') then
text = '^' .. replace(text, '([ |])', '%1^')
end
if find(content, 'infl=い') then
text = replace(text, 'しい', 'し.い')
end
if find(content, 'ja%-verb') then
text = replace(text, 'おう', 'お.う')
end
for parameter in itersplit(text, '|') do
if find(parameter, '[あ-ー]') then
table.insert(readings, parameter)
end
end
end
local function go()
for parameters in itermatch(content, '{{ja%-adj|([^}]+)}}') do
process(parameters)
end
for parameters in itermatch(content, '{{ja%-noun|([^}]+)}}') do
process(parameters)
end
for parameters in itermatch(content, '{{ja%-verb|([^}]+)}}') do
process(parameters)
end
for parameters in itermatch(content, '{{ja%-verb%-suru|([^}]+)}}') do
process(parameters)
end
for parameters in itermatch(content, '{{ja%-phrase|([^}]+)}}') do
process(parameters)
end
for parameters in itermatch(content, '{{ja%-pos|([^}]+)}}') do
process(parameters)
end
for parameters in itermatch(content, '{{ja%-altread|([^}]+)}}') do
process(parameters)
end
end
readings = {}
content = mw.title.new(pagename):getContent()
if content then
go()
else
return readings
end
for pagename in itermatch(content, '{{ja%-see|([^|}]+)') do
local readings_old = readings
content = mw.title.new(pagename):getContent()
if content then
go()
end
if #readings_old == #readings then
-- [[解れる]]→[[ほぐれる]], scanning [[ほぐれる]] will find nothing
table.insert(readings, pagename)
end
end
readings = require("Module:table").removeDuplicates(readings)
return readings
end
function export.r(frame)
local one = frame.args[1] or ''
local two = frame.args[2] or ''
local three = frame.args[3] or ''
local four = frame.args[4] or ''
local jp = ''
local tr = ''
local gloss = frame.args['gloss'] or ''
local choice = ''
if find(one, '[0-9]') then
choice = one
jp = two
linktitle = three
gloss = (gloss ~= '' and gloss or four)
elseif one == 'ja' then
choice = ''
jp = two
linktitle = three
gloss = (gloss ~= '' and gloss or four)
else
choice = ''
jp = one
linktitle = two
gloss = (gloss ~= '' and gloss or three)
end
local readings = export.extract_ja_readings(jp)
if #readings > 1 then
if choice ~= '' then
tr = readings[tonumber(choice)]
else
return '{{ja-r|' .. jp .. '|ーーーーー}}\n' .. require("Module:debug").highlight_dump(readings)
end
else
tr = (readings[1] and readings[1] or jp)
end
-- if term is pure kana and kana is identical
if replace(jp, '[あ-ー]', '') == '' and tr == jp then
tr = ''
end
if gloss ~= '' then
gloss = ': ' .. gloss
end
if tr ~= '' then
tr = '|' .. tr
end
if linktitle ~= '' then
jp = 'linkto=' .. jp .. '|' .. linktitle
end
return '{{ja-r|' .. jp .. tr .. '}}' .. gloss
--[[
変換済みの言葉を再変換
・選択してスペースキーを押す
・または選択してWin+Cを押す
]]
end
function export.hzreport(frame)
local text = {}
local candidates = mw.loadData('Module:User:Suzukaze-c/02/hz').hz
local rows = {}
local y, n = '✔️', '❌️' -- taking advantage of colored emoji. with vs16
for hz in itersplit(candidates[frame.args[1]], '') do
local content = mw.title.new(hz):getContent() or ''
local zh = find(content, '==Chinese==') and y or n
local def = find(content, '{{zh%-[a-z ]+}}\n\n#') and y or n
local der = find(content, '===Compounds===+\n{{zh%-der') and y or n
local uns = match(content, '|sim=(.)') or ''
local unt = match(content, '|tra=(.)') or ''
local ufs = match(content, '{{zh%-forms|s=(.)') or ''
local uft = match(content, '{{zh%-see|([^}]+)}}') or ''
local goh = find(content, '===Glyph origin===') and y or n
local histf = find(content, '{{[Hh]an[_ ]ety[ml]}}') and y or n
local ids = find(content, '|ids=[⿰⿱⿲⿳⿴⿵⿶⿷⿸⿹⿺⿻]') and y or n
local yue = match(content, '|c=([a-z0-9,]+)') or ''
local prc_tw = find(content, '|m=[㐀-鿕𠀀-𬺡]') and y or n
uft = replace(uft, "[{|']", '.')
if len(uft) > 6 then uft = sub(uft, 1, 5) .. '◆' end
if len(yue) > 6 then yue = sub(yue, 1, 5) .. '◆' end
hz = '[' .. tostring(mw.uri.canonicalUrl(hz,'action=edit')) .. ' ' .. hz .. ']'
local cells = { hz, zh, def, der, uns, unt, ufs, uft, goh, histf, ids, yue, prc_tw }
table.insert(rows, '| ' .. table.concat(cells, ' || '))
end
table.insert(text, '[[#wpTextbox1]]')
table.insert(text, '{| class="wikitable sortable Hani"')
table.insert(text, '! hz || zh? || def || der || unS || unT || ufS || ufT || goh || histF || ids || yue || prc/tw')
table.insert(text, '|-')
table.insert(text, table.concat(rows, '\n|-\n'))
table.insert(text, '\n|}')
return table.concat(text, '\n')
end
function export.newja(frame)
local a = frame.args[1] and frame.args or frame:getParent().args
local m_languages = require('Module:languages')
local m_scripts = require('Module:scripts')
local lang_ja = m_languages.getByCode('ja')
local sc_Jpan = m_scripts.getByCode('Jpan')
local sc_Hrkt = m_scripts.getByCode('Hrkt')
local sc_Hani = m_scripts.getByCode('Hani')
local pagename = a['testing_pagename'] or mw.title.getCurrentTitle().text
local header_level = 2 -- header level
local this_content = mw.title.new(mw.title.getCurrentTitle().text):getContent() or ''
local pos_datas = {
['a'] = {'Adjective:adj'},
['n'] = {'Noun:noun'},
['p'] = {'Phrase:phrase'},
['v'] = {'Verb:verb'},
['suru'] = {'Verb:verb-suru'},
['vf'] = {'Verb:verb form'},
['ad'] = {'Adverb:pos|adverb'},
['adv'] = {'Adverb:pos|adverb'},
['i'] = {'Interjection:pos|interjection'},
['pn'] = {'Proper noun:pos|proper noun'},
['af'] = {'Adjective:pos|adjective form'},
['c'] = {'Conjunction:pos|conjunction'},
['s'] = {'Noun:noun', 'Verb:verb-suru'},
['an'] = {'Adjective:adj', 'Noun:noun'},
}
local pos_aliases = {
[''] = 'n',
}
for pos_codes, array in pairs(pos_datas) do
for i, name_and_template in ipairs(array) do
name_and_template = split(name_and_template, ':')
pos_datas[pos_codes][i] = {
['name'] = name_and_template[1],
['template'] = name_and_template[2],
}
end
end
local verb_types_corresp = {
['1'] = '2', -- ichidan is type=2
['5'] = '1', -- godan is type=1
}
local etymology_magic_words = {
['der ([^.,{}\\]+)'] = function(a)
return '{{der|ja|' .. replace(a, '^([a-z]+) ', '%1|') .. '}}'
end,
['bor ([^.,{}\\]+)'] = function(a)
return '{{bor|ja|' .. replace(a, '^([a-z]+) ', '%1|') .. '}}'
end,
['b ([^.,{}\\]+)'] = function(a)
return '{{bor|ja|' .. replace(a, '^([a-z]+) ', '%1|') .. '}}'
end,
['obor ([^.,{}\\]+)'] = function(a)
return '{{obor|ja|' .. replace(a, '^([a-z]+) ', '%1|') .. '}}'
end,
['int'] = function(a)
return '{{internationalism|ja}}; see {{cog|en|}}'
end,
['rd ([a-z_]+) ([a-z_]+)'] = function(a, b)
return '{{rendaku2|' .. replace(a, '_', '̄') .. '|' .. replace(b, '_', '̄') .. '}}'
end,
['rfe (.+)'] = function(a)
return '{{rfe|ja' .. (a and '|' .. a or '') .. '}}'
end,
['ltc'] = function(a)
return '{{der|ja|ltc|-}} {{ltc-l|' .. pagename .. '}}'
end,
}
local usage_notes_magic_words = {
['bio'] = '{{U:ja:biology}}'
}
local output = {}
-- parameters sorted by appearance
local params = {
[2] = {list = true, allow_holes = true}, -- pos and def
['en'] = {type = 'number'}, -- etymology number
['e'] = {}, -- etymology text
['eid'] = {}, -- [[t:etymid]]
['l'] = {}, -- alt in header
['ate'] = {allow_empty = true}, -- ateji
['y'] = {allow_empty = true}, -- yomi
['r'] = {allow_empty = true}, -- rendaku
['ll'] = {}, -- alt in [[t:ja-kanjitab]]
['w'] = {allow_empty = true}, -- wikipedia
['wen'] = {allow_empty = true}, -- en.wikipedia
['file'] = {}, -- file
['caption'] = {}, -- file caption
['nop'] = {allow_empty = true}, -- no [[t:ja-pron]]
['a'] = {}, -- accent
['dev'] = {}, -- devoicing
['head'] = {}, -- head
[1] = {allow_empty = true, default = ''}, -- kana
['tr'] = {}, -- transivity
['ak'] = {}, -- alternate kana
['mw'] = {}, -- measure word/counter
['kyu'] = {}, -- kyuujitai
['hh'] = {}, -- historical hiragana
['un'] = {}, -- usage notes
['sy'] = {}, -- synonyms
['an'] = {}, -- antonyms
['hsup'] = {}, -- hypernyms
['hsub'] = {}, -- hyponyms. super- and sub- bc i'm big dumb
['co'] = {}, -- coordinate terms
['de'] = {}, -- derived terms
['re'] = {}, -- related terms
['desc'] = {}, -- descendants
['al'] = {}, -- see also
['reflist'] = {allow_empty = true}, -- references header
['c'] = {}, -- [[t:C]]
['cln'] = {}, -- [[t:cln]]
['testing_pagename'] = {},
-- `allow_empty = true`
-- instead of `type = 'boolean'`
-- `|r=` is sufficient for saying 'rendaku yes'
-- instead of `|r=y`
-- "typo" "correction" also takes up time
['ateji'] = {alias_of = 'ate'},
['yomi'] = {alias_of = 'y'},
['h'] = {alias_of = 'hh'},
['s'] = {alias_of = 'sy'},
['syn'] = {alias_of = 'sy'},
['ant'] = {alias_of = 'an'},
['d'] = {alias_of = 'de'},
['also'] = {alias_of = 'al'},
['cat'] = {alias_of = 'c'},
}
a = require('Module:parameters').process(a, params)
local function waapuro_to_kana(text)
if text == '' then return text end -- with just one parameter, [[mod:typing-aids]] will fallback to language 'all' and treat parameter 1 as the text to convert. for us that will return 'ja'
text = replace(text, '-', '@@@') -- preserve hyphen
text = replace(text, '_', '-') -- instead of hyphen, use underscore for chouonpu
--text = require('Module:typing-aids').replace({'ja', text})
text = replace(text, '@@@', '-') -- restore hyphen
text = replace(text, ' ', ' ') -- replace fullwidth space
return text
end
local function wr(text)
table.insert(output, text)
end
local function wrh(text)
wr('\n' .. mw.ustring.rep('=', header_level) .. text .. mw.ustring.rep('=', header_level))
end
local function hl(n)
header_level = header_level + n
end
local function needs_reading(pagename, kana)
return not (kana or find(pagename, '^[ぁ-ー ^%%.%[%]%-]+$'))
end
local function generate_tab_params(kana, a)
-- 銀行
-- gin,kou
-- gin|kou|yomi=o
-- 文字
-- mon;mo,ji
-- mon|k1=mo|ji|yomi=o
-- 送り仮名
-- [oku]ri[ka;ga][na]
-- oku|ri|ka|k3=ga|na|yomi=k
-- 送仮名
-- [oku:ri][ka;ga][na]
-- oku|o1=ri|ka|k3=ga|na|yomi=k
-- 満漢全席
-- man-kan zen,seki
-- man|kan|zen|seki|yomi=o
-- 犬
-- inu
-- inu
-- because pressing shift is effort
local yomi
local params = {}
if find(kana, '%[') then
yomi = 'k'
for yomigana in itermatch(kana, '%[(.-)%]') do
table.insert(params, yomigana)
end
elseif sc_Hani:countCharacters(pagename) > 0 then
if find(kana, ',') then
yomi = 'o'
end
kana = replace(kana, '%^', '')
kana = replace(kana, '[ .-]', ',')
if kana ~= '' then
params = split(kana, ',')
end
end
for i, yomigana in ipairs(params) do
yomigana = replace(yomigana, ';', '|k' .. i .. '=')
yomigana = replace(yomigana, ':', '|o' .. i .. '=')
params[i] = yomigana
end
for i, _ in ipairs(params) do
params[i] = require('Module:ja').kata_to_hira(params[i])
end
if a['r'] then
table.insert(params, 'r=y')
end
if a['ate'] then
table.insert(params, 'ateji=' .. (a['ate'] == '' and 'y' or a['ate']))
end
yomi = a['y'] or yomi
if yomi then
table.insert(params, 'yomi=' .. yomi)
end
if a['ll'] then
table.insert(params, 'alt=' .. replace(a['ll'], '、', ','))
end
return params
end
local function generate_links_list(text)
-- 3密
-- * {{ja-l|3密}}
-- 3密 sanmitu
-- * {{ja-r|3密|さんみつ}}
-- 3密 sanmitu,sense=Sense\q=Qualifier\三つの密 mittu no mitu\gloss=Gloss
-- * {{ja-r|3密|さんみつ}}
-- * {{sense|Sense}} {{q|Qualifier}} {{ja-r|三つの密|みっつ の みつ|gloss=Gloss}}
-- いい_ね
-- * {{ja-r|いい ね}}
local params = {}
for i, item in ipairs(split(text, '[,、]')) do
item = split(item, '[\\¥]')
local q
local sense
local output_link = {}
local main_found = false
local pagename, kana
local r_or_l
while #item > 0 do
if find(item[1], '^q=') then
q = replace(item[1], '^q=', '')
elseif find(item[1], '^sense=') then
sense = replace(item[1], '^sense=', '')
elseif not main_found then
pagename, kana = match(item[1], '^([^ ]-)[ ](.+)$')
pagename = pagename or item[1] -- if match() returns nil
pagename = (pagename and replace(pagename, '_', ' ') or pagename)
kana = (kana and replace(kana, ' ', ' ') or kana)
kana = (kana and waapuro_to_kana(kana) or kana)
if not kana then
kana = export.extract_ja_readings(pagename)
if #kana == 1 then
kana = kana[1]
else
kana = nil
end
end
r_or_l = needs_reading(pagename, kana) and 'l' or 'r'
table.insert(output_link, 'ja-' .. r_or_l)
table.insert(output_link, pagename)
table.insert(output_link, kana)
main_found = true
else
table.insert(output_link, item[1])
end
table.remove(item, 1)
end
table.insert(
params,
'* ' .. (sense and '{{sense|' .. sense .. '}} ' or '') .. (q and '{{q|' .. q .. '}} ' or '') .. '{{' .. table.concat(output_link, '|') .. '}}'
)
end
return params
end
local kana = a[1]
local kana_no_hyphens
-- convert kana from romaji to kana
kana = waapuro_to_kana(kana)
kana = mw.ustring.toNFC(kana) -- wtf?
-- convert fullwidth CJK symbols to halfwidth
kana = replace(kana, '(.)', {
['「'] = '[',
[';'] = ';',
[':'] = ':',
['」'] = ']',
['、'] = ',',
['。'] = '.',
[' '] = ' ',
})
-- generate [[t:ja-kanjitab]]
local tab
if a['ll'] or sc_Hani:countCharacters(pagename) > 0 then
local tab_params = generate_tab_params(kana, a)
tab_params = table.concat(tab_params, '|')
tab = '{{ja-kanjitab' .. (tab_params and '|' .. tab_params) .. '}}'
end
-- remove markup for generating [[t:ja-kanjitab]] from kana
kana = replace(kana, '[0-9]', '')
-- kun
kana = replace(kana, '(%[)([^%[%]]-):([^%[%]]-)(%])', '%1%2%4%3')
kana = replace(kana, '(%[)([^%[%]]-);([^%[%]]-)(%])', '%3')
kana = replace(kana, '(%[)([^%[%]]-)(%])', '%2')
-- on
kana = replace(kana, '([ .,-])([^ .,-]-;)', '%1')
kana = replace(kana, '^([^ .,-]-;)', '')
kana = replace(kana, ',', '')
-- for [[t:ja-pron]]
kana_no_hyphens = replace(kana, '[-^]', '')
-- blank if it's the same as the pagename. avoid unnecessary template input
if kana == pagename then kana = '' end
if kana_no_hyphens == pagename then kana_no_hyphens = '' end
-- automatic |head= using |1= (kana)
if replace(kana, '[ %^%-%.]+', '') == pagename and not a['head'] then
a['head'] = '[[' .. replace(kana, '([ %^%-%.]+)', ']]%1[[') .. ']]'
a['head'] = replace(a['head'], '%[%[([はがをにのもで])%]%]', '%1')
end
-- process etymology
-- process usage notes
if a['e'] then
-- expand shortcuts
for magic_word, f in pairs(etymology_magic_words) do
a['e'] = replace(a['e'], '\\' .. magic_word, f)
a['e'] = replace(a['e'], ' }}{{', '}} {{')
end
-- automatically fill empty {{bor}}/{{der}}/{{cog}}
if a[2][2] and match(a[2][2], '%[%[(.-)%]%]') then
a['e'] = replace(a['e'], '({{[a-z]+[+]?|ja|[a-z]+)(}})', function(template_head, template_tail) return template_head .. '|' .. match(a[2][2], '%[%[(.-)%]%]') .. template_tail end)
a['e'] = replace(a['e'], '(see {{cog|en|)(}})', function(template_head, template_tail) return template_head .. match(a[2][2], '%[%[(.-)%]%]') .. template_tail end)
end
-- add final period
if not find(a['e'], '%.$') and not find(a['e'], 'UNIQ.-QINU') and not find(a['e'], '{{rendaku2[^}]+}}$') and not find(a['e'], '{{pedia[^}]+}}$') and not find(a['e'], '%-%->$') and not find(a['e'], '{{rfe[^\n]+$') then
a['e'] = a['e'] .. '.'
end
-- add leading 'From'
if find(a['e'], '^{{compound|') then
a['e'] = 'From ' .. a['e']
end
-- change leading [[t:bor]], [[t:der]] to [[t:bor+]], [[t:der+]]
if find(a['e'], '^{{bor|') or find(a['e'], '^{{der|') then
a['e'] = replace(a['e'], '^{{bor|', '{{bor+|')
a['e'] = replace(a['e'], '^{{der|', '{{der+|')
end
end
if a['un'] then
-- expand shortcuts
a['un'] = replace(a['un'], '\\([a-z]+)', usage_notes_magic_words)
end
-- automatically make an {{etymid}}
if a['en'] and not a['eid'] then
if match(pagename, '^[㐀-鿕]+$') then
a['eid'] = kana_no_hyphens
elseif a['ll'] then
a['eid'] = match(a['ll'], '^([^、,]+)')
end
end
-- write Japanese header
-- write etymology section
-- write etymology
-- write [[t:wikipedia]]
-- write [[t:ja-kanjitab]]
-- write alternative forms
if match(this_content, 'ja%-readings') and not a['en'] then
a['en'] = 0
end
if a['en'] then
-- en = 0
-- for adding new sections under a single character entry like [[字]], where you want an Etymology section for sanity, but just one, and no Japanese header
if a['en'] == 1 and not match(this_content, 'ja%-readings') then wrh('Japanese') end
hl(1)
wrh('Etymology' .. (a['en'] == 0 and '' or ' ' .. a['en']))
wr(a['eid'] and ('{{etymid|ja|' .. a['eid'] .. '}}') or nil)
hl(1)
wr(tab and tab or nil)
wr(a['w'] and ('{{wikipedia|lang=ja' .. (a['w'] == '' and '' or '|' .. a['w']) .. '}}') or nil)
wr(a['wen'] and ('{{wikipedia|lang=en|' .. a['wen'] .. '}}') or nil)
wr(a['file'] == 'rfi' and '{{rfi|ja}}' or nil)
if a['file'] == 'rfi' then a['file'] = nil end
wr(a['file'] and ('[[File:' .. a['file'] .. '|thumb|right' .. (a['caption'] and '|' .. a['caption'] or '').. ']]') or nil)
wr(a['e'])
if a['l'] then
wrh('Alternative forms')
wr(table.concat(generate_links_list(a['l']), '\n'))
end
else
wrh('Japanese')
hl(1)
wr(tab and tab or nil)
wr(a['w'] and ('{{wikipedia|lang=ja' .. (a['w'] == '' and '' or '|' .. a['w']) .. '}}') or nil)
wr(a['wen'] and ('{{wikipedia|lang=en|' .. a['wen'] .. '}}') or nil)
wr(a['file'] == 'rfi' and '{{rfi|ja}}' or nil)
if a['file'] == 'rfi' then a['file'] = nil end
wr(a['file'] and ('[[File:' .. a['file'] .. '|thumb|right' .. (a['caption'] and '|' .. a['caption'] or '').. ']]') or nil)
if a['l'] then
wrh('Alternative forms')
wr(table.concat(generate_links_list(a['l']), '\n'))
end
if a['e'] then
wrh('Etymology')
wr(a['e'])
end
end
if sc_Hani:countCharacters(pagename) > 0 and not a[2][1] and not a[2][2] then
-- if title has kanji and not given pos or def
-- write [[t:ja-see]] by combining title and reading
if a['e'] then
wrh('Definitions')
end
pagename_to_kana = pagename
for char in itermatch(replace(replace(a[1], '[::]', ''), '[ぁ-ン]+[;;]', ''), '[ぁ-ン]+') do
pagename_to_kana = replace(pagename_to_kana, '[㐀-龥]', char, 1)
end
wr('{{ja-see|' .. pagename_to_kana .. '}}')
elseif a[2][1] and a[2][1] ~= '' and not find(a[2][1], '[a-z]') then
-- if pos is non-latin
-- write [[t:ja-see]] using that text
if a['e'] then
wrh('Definitions')
end
wr('{{ja-see|' .. table.concat(a[2], '|') .. '}}')
else
-- write [[t:ja-pron]]
if not a['nop'] then
-- 1DJR,2,3-
-- 0DJR NHK
local pron_params = {}
if kana_no_hyphens ~= '' or needs_reading(pagename, kana) then
table.insert(pron_params, kana_no_hyphens)
end
if a['a'] then
for i, acc_item in ipairs(split(a['a'], '[,、]')) do
local acc, acc_ref = match(acc_item, '^(%d+)(.*)$')
acc_ref = replace(acc_ref, ' ', ',')
acc_ref = string.upper(acc_ref)
if acc_ref == '' then
acc_ref = 'DJR'
elseif acc_ref == '-' then
acc_ref = nil
end
if acc_ref and not a['reflist'] then
a['reflist'] = true
end
table.insert(pron_params, 'acc' .. (i > 1 and i or '') .. '=' .. acc)
table.insert(pron_params, acc_ref and 'acc' .. (i > 1 and i or '') .. '_ref=' .. acc_ref or nil)
table.insert(pron_params, a['dev'] and 'dev=' .. a['dev'] or nil)
end
end
wrh('Pronunciation')
wr('{{ja-pron' .. (#pron_params > 0 and '|' .. table.concat(pron_params, '|') or '') .. '}}')
end
-- if no pos or def parameters
-- then generate a default Noun and [[t:rfdef]]
if a[2]['maxindex'] == 0 then
a[2] = {''}
a[2]['maxindex'] = 1
end
i = 1
while i <= a[2]['maxindex'] do
-- 犬 <empty string>
-- 犬 n
-- 赤い a
-- 赤い a,i
-- 明らか a,na
-- 画然 a,tari
-- 異常 an
-- 食べる v,2
local pos_code
local defs
local type, infl
pos_code = a[2][i] or ''
defs = {
a[2][i + 1] or '{{rfdef|ja}}',
a[2][i + 2] or '{{rfdef|ja}}',
}
-- shortcut for {{lb}}
-- [slang,dated]
defs[1] = replace(defs[1], '^%[([^%[%]]+)%]', function(labels) return '{{lb|ja|' .. replace(labels, ',', '|') .. '}}' end)
defs[1] = replace(defs[1], '# %[([^%[%]]+)%]', function(labels) return '# {{lb|ja|' .. replace(labels, ',', '|') .. '}}' end)
defs[2] = replace(defs[2], '^%[([^%[%]]+)%]', function(labels) return '{{lb|ja|' .. replace(labels, ',', '|') .. '}}' end)
defs[2] = replace(defs[2], '# %[([^%[%]]+)%]', function(labels) return '# {{lb|ja|' .. replace(labels, ',', '|') .. '}}' end)
local match_a, match_b = match(pos_code, '^(.+),(.+)$')
if match_a then pos_code, type = match_a, match_b end
pos_code = pos_aliases[pos_code] or pos_code
if pos_code == 'v' and verb_types_corresp[type] then
type = verb_types_corresp[type]
end
-- default type
if not type then
if pos_code == 'an' then
type = 'na'
elseif pos_code == 'v' then
type = '1'
elseif pos_code == 'a' then
type = 'i'
end
end
-- adjectives use infl
if pos_code == 'an' or pos_code == 'a' then
infl = type
type = nil
end
-- suffixes are uhhh. uh. fuck it it's all 活用 anyway
if pos_code == 'suffix' then
if type == 'i' then
infl = type
type = nil
end
end
-- get data
local pos_data = pos_datas[pos_code]
-- create fallback data
pos_data = pos_data or {
{
['name'] = replace(pos_code, '^.', mw.ustring.upper),
['template'] = 'pos|' .. pos_code,
},
}
-- write header, etc
for ii, name_and_template in ipairs(pos_data) do
local is_a_or_v = false
name = name_and_template['name']
template = name_and_template['template']
is_a_or_v = match(template, '^adj') or match(template, '^verb') or match(template, '^pos|suffix') -- prevent inappropriate addition of parameters (`noun|infl=na`)
if (sc_Hani:countCharacters(pagename) > 0) then
template = template .. (a['head'] and '|head=' .. a['head'] or '')
template = template .. (kana ~= '' and '|' .. kana or '')
else
template = template .. (a['head'] and '|' .. a['head'] or (kana ~= '' and '|' .. kana or ''))
end
template = template .. (a['ak'] and '|' .. replace(waapuro_to_kana(a['ak']), ',', '|') or '')
if is_a_or_v then
template = template .. (a['tr'] and '|tr=' .. a['tr'] or '')
template = template .. (type and '|type=' .. type or '')
template = template .. (infl and '|infl=' .. infl or '')
end
template = template .. (a['mw'] and '|count=' .. a['mw'] or '')
template = template .. (a['kyu'] and '|kyu=' .. a['kyu'] or '')
template = template .. (a['hh'] and '|hhira=' .. waapuro_to_kana(a['hh']) or '')
wrh(name)
wr('{{ja-' .. template .. '}}')
wr('')
wr('# ' .. defs[ii])
if is_a_or_v then
local kana_stem = sub(kana ~= '' and kana or pagename, 0, -2)
kana_stem = (kana == '' and kana or kana_stem) -- the templates will be smart if you do not give it a reading
local kana_last = sub(kana ~= '' and kana or pagename, -1)
kana_last = lang_ja:transliterate(kana_last, sc_Hrkt)
if type or pos_code == 's' or pos_code == 'suru' then
hl(1)
wrh('Conjugation')
if type == '2' then
wr('{{ja-ichi' .. (kana_stem ~= '' and '|' .. kana_stem or '') .. '}}')
elseif pos_code == 's'or pos_code == 'suru' then
wr('{{ja-suru' .. (kana ~= '' and '|' .. kana or '') .. '}}')
elseif type == '1' then
wr('{{ja-go-' .. kana_last .. (kana_stem ~= '' and '|' .. kana_stem or '') .. '}}')
else
wr('{{rfinfl|ja}}')
end
hl(-1)
end
if infl then
hl(1)
wrh('Inflection')
if infl == 'na' then
wr('{{ja-' .. infl .. (kana ~= '' and '|' .. kana or '') .. '}}')
elseif infl == 'shiku' or infl == 'ku' then
-- ja-conj-bungo will not be smart
wr('{{ja-conj-bungo|' .. sub(kana ~= '' and kana or pagename, 0, -2) .. '|adj-' .. infl .. '}}')
elseif infl == 'i' then
wr('{{ja-' .. infl .. (kana_stem ~= '' and '|' .. kana_stem or '') .. '}}')
else
wr('{{rfinfl|ja}}')
end
hl(-1)
end
end
if i == 1 and ii == 1 then
hl(1)
if a['un'] then
wrh('Usage notes')
wr('* ' .. a['un'])
end
if a['sy'] then
wrh('Synonyms')
wr(table.concat(generate_links_list(a['sy']), '\n'))
end
if a['an'] then
wrh('Antonyms')
wr(table.concat(generate_links_list(a['an']), '\n'))
end
if a['hsup'] then
wrh('Hypernyms')
wr(table.concat(generate_links_list(a['hsup']), '\n'))
end
if a['hsub'] then
wrh('Hyponyms')
wr(table.concat(generate_links_list(a['hsub']), '\n'))
end
if a['co'] then
wrh('Coordinate terms')
wr('*' .. replace(table.concat(generate_links_list(a['co']), ','), '%*', ''))
end
if a['de'] then
wrh('Derived terms')
wr(table.concat(generate_links_list(a['de']), '\n'))
end
if a['re'] then
wrh('Related terms')
wr(table.concat(generate_links_list(a['re']), '\n'))
end
if a['desc'] then
wrh('Descendants')
wr('* ' .. a['desc'])
end
if a['al'] then
wrh('See also')
wr(table.concat(generate_links_list(a['al']), '\n'))
end
hl(-1)
end
end
-- advance i by the number of PoS headers produced
i = i + 1 + (#pos_data)
end
end
if (a['reflist'] or (a['e'] and find(a['e'], 'UNIQ.-QINU'))) and (a['reflist'] ~= 'n') and (not a['en']) then
if a['en'] then
hl(-1)
end
wrh('References')
wr('<references />')
end
if a['c'] or a['cln'] then
wr('')
if a['c'] then
a['c'] = replace(a['c'], '^.', mw.ustring.upper)
a['c'] = replace(a['c'], '(,)(.)', function(a, b) return a .. mw.ustring.upper(b) end)
wr('{{C|ja|' .. replace(a['c'], '[,]', '|') .. '}}')
end
if a['cln'] then
wr('{{cln|ja|' .. replace(a['cln'], '[,]', '|') .. '}}')
end
end
output = table.concat(output, '\n')
-- html comments
-- real html comments are discarded before the module ever knows
output = replace(output, '<%-%-', '<!--')
-- trim: otherwise `Etymology n` headers create a leading newline
return trim(output)
end
function export.newhz(frame)
local m_zh = require("Module:zh")
local m_zh_new = require("Module:zh-new")
local a = frame.args
local character = mw.title.getCurrentTitle().text
local en = tonumber(a['e']) or 0 -- etym_number
local hl = 3 -- header_level
local function head(text) return '\n' .. mw.ustring.rep('=', hl) .. text .. mw.ustring.rep('=', hl) end
local function hcr(a) return (a > 0 and hl + 1 or hl - 1) end -- header_{in|de}crement
local content = mw.title.new(character):getContent() or ''
local simp_form = a['s'] or match(content, '|sim=(.)') or match(content, '|s=(.)') or false
--local trad_form = a['t'] or match(content, '|tra=(.)') or match(content, '|t=(.)') or false
local alt_form = a['alt'] or a['v'] or match(content, '|alt=([^|}]+)') or false
local zhwp_data = require('Module:User:Suzukaze-c/02/zhwp')
-- ----
local add_han_etym = false
local candidates = mw.loadData('Module:User:Suzukaze-c/02/hz').hz
if en <= 1 then
if find(candidates['hanetym'], character) or require("module:zh-glyph/phonetic/list")[character] then
add_han_etym = true
end
end
local zh_see
if not a['ignore'] then
if trad_form or m_zh.ts_determ(character) == 'simp' then
zh_see = (trad_form or m_zh.st(character))
end
end
if a[1] and match(a[1], '^[㐀-鿕𠀀-𬺡]$') then
zh_see = a[1] .. (a[2] and '|' .. a[2] or '')
end
local zh_forms = ''
if simp_form or m_zh.ts_determ(character) == 'trad' then
zh_forms = zh_forms .. '|s=' .. (simp_form or m_zh.ts(character))
end
if alt_form then
zh_forms = zh_forms .. '|alt=' .. alt_form
end
local zh_wp
if a['wp'] then
if a['wp'] == 'y' then
zh_wp = ''
elseif a['wp'] == '' then
zh_wp = false
else
zh_wp = '|' .. replace(a['wp'], ',', '|')
end
end
if zhwp_data.single_char_title[character] or zhwp_data.contains_astral[character] or zhwp_data.single_char_title[m_zh.ts(character)] then
zh_wp = ''
end
-- ----
local output = {}
local function write(text) table.insert(output, text) end
-- ----
if en <= 1 then
write('==Chinese==')
end
if en == 0 then
if not zh_see then write('{{zh-forms' .. zh_forms .. '}}') end
if zh_wp then
write('{{zh-wp' .. zh_wp .. '}}')
end
end
if a['ge'] or add_han_etym then
write(head('Glyph origin'))
if add_han_etym then write('{{Han etym}}') end
if a['ge'] then write(a['ge']) end
if zh_see and en == 0 then
write(head('Definitions'))
end
end
if en > 0 then
write(head('Etymology ' .. tostring(en)))
if not zh_see then write('{{zh-forms' .. zh_forms .. '}}') end
if zh_wp then
write('{{zh-wp' .. zh_wp .. '}}')
end
end
if en > 0 then
hl = hcr(1)
end
if zh_see then
write('{{zh-see|' .. zh_see .. '}}')
else
write(head('Pronunciation'))
local m, c, h, mn, w = a['m'] or false, a['c'] or false, a['h'] or false, a['mn'] or false, a['w'] or false
local mc, oc = a['mc'] or false, a['oc'] or false
if m and find(m, '[ㄅ-ㄩ]') then
m = replace(m, '[^.ˊˇˋㄅ-ㄩ]', '@')
m = replace(m, '@+', '@')
m = replace(m, '^@+', '')
m = replace(m, '@+$', '')
m = replace(m, '@', ',')
m = require("module:cmn-pron").zhuyin_py(m)
end
if m and find(m, '[0-9]') then
m = replace(m, '1', '̄')
m = replace(m, '2', '́')
m = replace(m, '3', '̌')
m = replace(m, '4', '̀')
m = replace(m, 'v', 'ü')
end
if not m then
if require("module:zh/data/cmn-tag").MT[character] then -- if there is cmn-tag data
m = character
else
m = mw.ustring.gsub(m_zh_new.pytemp(character,'','',''), ',', ', ') or false -- based on line from zh-new
if m == character or m == simp_form or m == m_zh.ts(character) then
-- pinyin conversion failed
m = false
end
end
end
if not c then
c = m_zh.check_pron(character, 'yue', 1) or false
if c and find(c, ',') then a['rhk'] = 'yes' end
end
if h and find(h, '[0-9]') then
h = replace(h, 'w', 'ṳ')
h = replace(h, '24', '̂')
h = replace(h, '11', '̀')
h = replace(h, '31', '́')
h = replace(h, '55', '')
h = replace(h, '2', '')
h = replace(h, '5', '̍')
end
if not h then
h = m_zh.check_pron(character, 'hak', 1) or false
end
if not mn then
mn = m_zh.check_pron(character, 'nan-hbl', 1) or false
end
if w and find(w, '[PSQR]') then
w = export.test_3(w)
end
if (not mc) and (mw.title.new('Module:zh/data/ltc-pron/' .. character).exists) then
mc = 'y'
end
if (not oc) and (mw.title.new('Module:zh/data/och-pron-BS/' .. character).exists or mw.title.new('Module:zh/data/och-pron-ZS/' .. character).exists) then
oc = 'y'
end
if not m then write('<!--') end
write('{{zh-pron')
if m then write('|m=' .. m .. (a['er'] and ',er=y' or '')) else write('|m=') end
if (a['ms'] or a['m-s']) then write('|m-s=' .. (a['ms'] or a['m-s'])) end
if c then write('|c=' .. c) end
if (a['ct'] or a['c-t']) then write('|c-t=' .. (a['ct'] or a['c-t'])) end
if a['g'] then write('|g=' .. a['g']) end
if h then write('|h=pfs=' .. h) end
if a['j'] then write('|j=' .. a['j']) end
if a['md'] then write('|md=' .. a['md']) end
if mn then write('|mn=' .. mn) end
if (a['mnn'] or a['mn_note']) then write('|mn_note=' .. (a['mnn'] or a['mn_note'])) end
if (a['mnt'] or a['mn-t']) then write('|mn-t=' .. (a['mnt'] or a['mn-t'])) end
if (a['mntn'] or a['mn-t_note']) then write('|mn-t_note=' .. (a['mntn'] or a['mn-t_note'])) end
if w then write('|w=' .. w) end
if a['x'] then write('|x=' .. a['x']) end
if mc or oc then
write('|mc=' .. (mc or ''))
write('|oc=' .. (oc or ''))
end
if a['ma'] then write('|ma=' .. a['ma']) end
write('|cat=' .. (a['cat'] or ''))
write('}}')
if not m then write('-->') end
if not m then write('{{rfp|cmn|Mandarin}}') end
write(head('Definitions'))
write('{{head|zh|hanzi}}')
write('')
if a['rcns'] then
write('# {{lb|zh|Taiwan}} {{n-g|Only used in personal names.}}')
else
write('# ' .. (a[1] or '{{rfdef|zh}}'))
end
if not a['noder'] then -- Lua error: not enough memory
local der = ''
local der_add = ''
if a['der'] then
der_add = a['der']
der_add = replace(der_add, '[^㐀-鿕𠀀-𬺡]+', '|') -- any non-hanzi text becomes separator
der_add = replace(der_add, '|+', '|')
der_add = replace(der_add, '^|', '')
der_add = replace(der_add, '|$', '')
der_add = '|' .. der_add
end
if match(character, '[𠀀-𬺡]') then
for title, _ in pairs(zhwp_data.contains_astral) do
if len(title) > 1 and match(title, character) then
der_add = der_add .. '|' .. title
end
end
end
der = frame:preprocess('{{subst:zh-der/new' .. (der_add or '') .. ((m and not match(m, ',')) and '|p=' .. mw.ustring.toNFC(m) or '') .. '}}')
if match(der, 'memory') then
write(head('Compounds'))
write('{{su#bst:zh-der/new' .. (der_add or '') .. (m and '|p=' .. replace(m, ',.+', '') or '') .. '}}') -- let you add zh-der in a separate edit in case Lua returns "out of memory"
elseif match(der, '[㐀-鿕𠀀-𬺡]') then
write(head('Compounds'))
write(der)
end
end
if a['also'] then
write(head('See also'))
write('* {{zh-l|' .. a['also'] .. '}}')
end
end
if en > 0 then
hl = hcr(-1)
end
if a['rtw'] or a['rhk'] or a['ryt'] or a['riso'] or a['rcns'] or a['rnan'] then
write(head('References'))
if a['rtw'] then write('* {{R:twedu|' .. a['rtw'] .. '}}') end
if a['rhk'] then write('* {{R:yue:mfccd}}') end
if a['ryt'] then write('* {{R:yue:jyut.net}}') end
if a['riso'] then write('* {{R:yue:Jyutping Database}}') end
if a['rcns'] then write('* {{R:zh:CNS|' .. replace(a['rcns'], '%-', '|') .. '}}') end
if a['rnan'] then
if find(a['rnan'], '%d') then
write('* {{R:nan:thcwd|' .. a['rnan'] .. '}}')
else
write('* {{R:nan:thcwdq}}')
end
end
end
return trim(table.concat(output, '\n'))
end
function export.newhzmul(frame)
local text = frame.args[3]
local char = mw.title.getCurrentTitle().text
local x = mw.title.new(char):getContent() or ''
local model = trim([==[
{{character info}}
==Translingual==
===Han character===
{{Han char|rn=$rs1|rad=$rad|as=$rs2|sn=$TotalStrokes|four=$FourCornerCode$four|canj=$Cangjie$canj|ids=$ids}}
# $Definition
====References====
{{Han ref|kx=$IRGKangXi|dkj=$IRGDaiKanwaZiten|dj=$IRGDaeJaweon|hdz=$IRGHanyuDaZidian|uh=$hex}}
]==])
local corr = {'一','丨','丶','丿','乙','亅','二','亠','人','儿','入','八','冂','冖','冫','几','凵','刀','力','勹','匕','匚','匸','十','卜','卩','厂','厶','又','口','囗','土','士','夂','夊','夕','大','女','子','宀','寸','小','尢','尸','屮','山','巛','工','己','巾','干','幺','广','廴','廾','弋','弓','彐','彡','彳','心','戈','戶','手','支','攴','文','斗','斤','方','无','日','曰','月','木','欠','止','歹','殳','毋','比','毛','氏','气','水','火','爪','父','爻','爿','片','牙','牛','犬','玄','玉','瓜','瓦','甘','生','用','田','疋','疒','癶','白','皮','皿','目','矛','矢','石','示','禸','禾','穴','立','竹','米','糸','缶','网','羊','羽','老','而','耒','耳','聿','肉','臣','自','至','臼','舌','舛','舟','艮','色','艸','虍','虫','血','行','衣','襾','見','角','言','谷','豆','豕','豸','貝','赤','走','足','身','車','辛','辰','辵','邑','酉','釆','里','金','長','門','阜','隶','隹','雨','靑','非','面','革','韋','韭','音','頁','風','飛','食','首','香','馬','骨','高','髟','鬥','鬯','鬲','鬼','魚','鳥','鹵','鹿','麥','麻','黃','黍','黑','黹','黽','鼎','鼓','鼠','鼻','齊','齒','龍','龜','龠'}
local corr_s = {['言']='讠',['門']='门',['食']='饣',['飛']='飞',['馬']='马',['見']='见',['貝']='贝',['糸']='纟',['車']='车',['長']='长',['韋']='韦',['風']='风',['金']='钅',['鳥']='鸟',['龍']='龙',['頁']='页',['齊']='齐',['麥']='麦',['龜']='龟',['魚']='鱼',['黽']='黾',['齒']='齿',['鹵']='卤'}
-- from text
local targets = {'RSUnicode','TotalStrokes','FourCornerCode','Cangjie','IRGKangXi','IRGDaiKanwaZiten','IRGDaeJaweon','IRGHanyuDaZidian','Definition'}
for _, property in ipairs(targets) do
local value = trim(match(text, 'k'..property..'%s+([^\n]+)') or '')
mw.log(property .. '|' .. value)
if property == 'RSUnicode' then
value = replace(value, ' .+', '') -- 龽
value = split(value, '%.')
model = replace(model, '$rs1', value[1])
model = replace(model, '$rs2', value[2])
elseif property == 'Definition' then
if value == '' or find(frame.args[1], 'x') then
model = replace(model, '# $Definition\n\n', '')
else
value = replace(value, ';', '\n#')
model = replace(model, '$Definition', value)
end
else
model = replace(model, '$'..property, value)
end
if property == 'FourCornerCode' and value ~= '' then
model = replace(model, '$four', '')
end
if property == 'Cangjie' and value ~= '' then
model = replace(model, '$canj', '')
end
end
-- read from existing page or manually provided
local ex = {'ids','four','canj'}
for _, property in ipairs(ex) do
model = replace(model, '$'..property, match(x, '|'..property..'=([^|}]+)') or frame.args[property] or '')
end
-- |rad=
model = replace(model, '(|rn=)(%d+)(\'?)(|rad=)($rad)(|)', function(a,b,c,d,e,f)
local z = corr[tonumber(b)]
return a .. b .. d .. (c == '\'' and corr_s[z] or z) .. f
end
)
-- remove empty dict fields
local template_ref_fields = {'kx','dkj','dj','hdz'}
for _, property in ipairs(template_ref_fields) do
model = replace(model, '|'..property..'=|', '|')
end
model = replace(model, '$hex', match(text, 'Unihan data for U.(%x+)'))
return model
end
function export.test_2(frame)
local search_base = 'https://en.wiktionary.org/w/index.php?title=Special%3ASearch&profile=default&fulltext=Search&search='
return '[' .. search_base .. mw.uri.encode(mw.text.unstripNoWiki(trim(frame.args[1])), 'PATH') .. ' ' .. trim(frame.args[2] or frame.args[1]) .. ']'
end
function export.test_3(text)
if type(text) == 'table' then text = text.args[1] end
local syllable, tone = match(text, '(.+)([PSQR])')
local voiced = false
if match(syllable, "^[bvdnlzg]") or match(syllable, "^m[^m]") or match(syllable, "jj") or match(syllable, "xx") or match(syllable, "hh") then
voiced = true
end
if match(syllable, '^[ctsjszh]+i') then
syllable = replace(syllable, '^[ctsjszh]+', { ['c']='j', ['ts']='j', ['ch']='q', ['tsh']='q', ['j']='jj', ['s']='x', ['sh']='x', ['z']='xx', ['zh']='xx' } )
end
syllable = replace(syllable, 'h$', 'q')
-- too lazy for vowels
if tone == 'P' and voiced then
tone = '3'
elseif tone == 'P' and not voiced then
tone = '1'
elseif tone == 'S' and voiced then
tone = '3'
elseif tone == 'S' and not voiced then
tone = '2'
elseif tone == 'Q' and voiced then
tone = '3'
elseif tone == 'Q' and not voiced then
tone = '2'
elseif tone == 'R' and voiced then
tone = '5'
elseif tone == 'R' and not voiced then
tone = '4'
end
return tone .. syllable
end
function export.test_9()
return '{{attn|ltc}}{{attn|och|Middle+Old Chinese needs to be distributed}}'
end
function export.test_10(content, target_header, pagename)
local section = 0
for header in itermatch(content, '==+([^\n=]+)==+\n') do
section = section + 1
if header == target_header then
break
end
end
return tostring(mw.uri.canonicalUrl(pagename,'action=edit§ion=' .. section))
end
function export.test_11()
local ciout = {}
local ziout = {}
local levels = {'Beginning','Elementary','Intermediate','Advanced'}
levels = {'Elementary','Intermediate'}
for _, level in ipairs(levels) do
table.insert(ciout, '*' .. level .. '\n**')
table.insert(ziout, '*' .. level .. '\n**')
local apcontent = mw.title.new('Appendix:HSK list of Mandarin words/' .. level .. ' Mandarin'):getContent()
apcontent = replace(apcontent, '{{l|cmn|([^|]+)|tr={{l|cmn|([^|]+)}}', '%2')
apcontent = replace(apcontent, '{{zh.l|([^/|]+)/([^/|]+)|', '%1')
apcontent = replace(apcontent, '{{zh.l|([^/|]+)/([^/|]+)/([^/|]+)|', '%1+%2')
apcontent = replace(apcontent, '%[%[([^%[%]]+)%]%] %(%[%[([^%[%]]+)%]%],', '%2')
apcontent = replace(apcontent, 'is called a [^\n]+', '')
for ci in itermatch(apcontent, '[㐀-鿕…]+') do
local cicontent = mw.title.new(ci):getContent() or ''
local ok = match(cicontent, 'zh%-pron')
local url = export.test_10(cicontent, 'Chinese', ci)
if not find(cicontent, level) then
table.insert(len(ci)==1 and ziout or ciout,
(ok and '' or '<mark>')
..
'[' .. url .. ' ' .. ci .. '], '
..
(ok and '' or '</mark>')
)
end
end
table.insert(ciout, '\n')
table.insert(ziout, '\n')
end
return table.concat(ciout, '')..table.concat(ziout, '')
end
function export.test_12(frame)
local out = {}
local kj = {
[1] = '亜哀挨愛曖悪握圧扱宛嵐安案暗以衣位囲医依委威為畏胃尉異移萎偉椅彙意違維慰遺緯域育一壱逸茨芋引印因咽姻員院淫陰飲隠韻右宇羽雨唄鬱畝浦運雲永泳英映栄営詠影鋭衛易疫益液駅悦越謁閲円延沿炎怨宴媛援園煙猿遠鉛塩演縁艶汚王凹央応往押旺欧殴桜翁奥横岡屋億憶臆虞乙俺卸音恩温穏下化火加可仮何花佳価果河苛科架夏家荷華菓貨渦過嫁暇禍靴寡歌箇稼課蚊牙瓦我画芽賀雅餓介回灰会快戒改怪拐悔海界皆械絵開階塊楷解潰壊懐諧貝外劾害崖涯街慨蓋該概骸垣柿各角拡革格核殻郭覚較隔閣確獲嚇穫学岳楽額顎掛潟括活喝渇割葛滑褐轄且株釜鎌刈干刊甘汗缶完肝官冠巻看陥乾勘患貫寒喚堪換敢棺款間閑勧寛幹感漢慣管関歓監緩憾還館環簡観韓艦鑑丸含岸岩玩眼頑顔願企伎危机気岐希忌汽奇祈季紀軌既記起飢鬼帰基寄規亀喜幾揮期棋貴棄毀旗器畿輝機騎技宜偽欺義疑儀戯擬犠議菊吉喫詰却客脚逆虐九久及弓丘旧休吸朽臼求究泣急級糾宮救球給嗅窮牛去',
[2] = '巨居拒拠挙虚許距魚御漁凶共叫狂京享供協況峡挟狭恐恭胸脅強教郷境橋矯鏡競響驚仰暁業凝曲局極玉巾斤均近金菌勤琴筋僅禁緊錦謹襟吟銀区句苦駆具惧愚空偶遇隅串屈掘窟熊繰君訓勲薫軍郡群兄刑形系径茎係型契計恵啓掲渓経蛍敬景軽傾携継詣慶憬稽憩警鶏芸迎鯨隙劇撃激桁欠穴血決結傑潔月犬件見券肩建研県倹兼剣拳軒健険圏堅検嫌献絹遣権憲賢謙鍵繭顕験懸元幻玄言弦限原現舷減源厳己戸古呼固股虎孤弧故枯個庫湖雇誇鼓錮顧五互午呉後娯悟碁語誤護口工公勾孔功巧広甲交光向后好江考行坑孝抗攻更効幸拘肯侯厚恒洪皇紅荒郊香候校耕航貢降高康控梗黄喉慌港硬絞項溝鉱構綱酵稿興衡鋼講購乞号合拷剛傲豪克告谷刻国黒穀酷獄骨駒込頃今困昆恨根婚混痕紺魂墾懇左佐沙査砂唆差詐鎖座挫才再災妻采砕宰栽彩採済祭斎細菜最裁債催',
[3] = '塞歳載際埼在材剤財罪崎作削昨柵索策酢搾錯咲冊札刷刹拶殺察撮擦雑皿三山参桟蚕惨産傘散算酸賛残斬暫士子支止氏仕史司四市矢旨死糸至伺志私使刺始姉枝祉肢姿思指施師恣紙脂視紫詞歯嗣試詩資飼誌雌摯賜諮示字寺次耳自似児事侍治持時滋慈辞磁餌璽鹿式識軸七叱失室疾執湿嫉漆質実芝写社車舎者射捨赦斜煮遮謝邪蛇勺尺借酌釈爵若弱寂手主守朱取狩首殊珠酒腫種趣寿受呪授需儒樹収囚州舟秀周宗拾秋臭修袖終羞習週就衆集愁酬醜蹴襲十汁充住柔重従渋銃獣縦叔祝宿淑粛縮塾熟出述術俊春瞬旬巡盾准殉純循順準潤遵処初所書庶暑署緒諸女如助序叙徐除小升少召匠床抄肖尚招承昇松沼昭宵将消症祥称笑唱商渉章紹訟勝掌晶焼焦硝粧詔証象傷奨照詳彰障憧衝賞償礁鐘上丈冗条状乗城浄剰常情場畳蒸縄壌嬢錠譲醸色拭食植殖飾触嘱織職辱',
[4] = '尻心申伸臣芯身辛侵信津神唇娠振浸真針深紳進森診寝慎新審震薪親人刃仁尽迅甚陣尋腎須図水吹垂炊帥粋衰推酔遂睡穂錘随髄枢崇数据杉裾寸瀬是井世正生成西声制姓征性青斉政星牲省凄逝清盛婿晴勢聖誠精製誓静請整醒税夕斥石赤昔析席脊隻惜戚責跡積績籍切折拙窃接設雪摂節説舌絶千川仙占先宣専泉浅洗染扇栓旋船戦煎羨腺詮践箋銭銑潜線遷選薦繊鮮全前善然禅漸膳繕狙阻祖租素措粗組疎訴塑遡礎双壮早争走奏相荘草送倉捜挿桑巣掃曹曽爽窓創喪痩葬装僧想層総遭槽踪操燥霜騒藻造像増憎蔵贈臓即束足促則息捉速側測俗族属賊続卒率存村孫尊損遜他多汰打妥唾堕惰駄太対体耐待怠胎退帯泰堆袋逮替貸隊滞態戴大代台第題滝宅択沢卓拓託濯諾濁但達脱奪棚誰丹旦担単炭胆探淡短嘆端綻誕鍛団男段断弾暖談壇地池知値恥致遅痴稚置緻竹畜逐蓄築秩窒茶着嫡中仲虫沖宙忠抽注昼柱衷酎鋳駐著貯丁弔庁兆町長挑帳張彫眺釣頂鳥朝脹貼超腸跳徴嘲潮澄調聴懲直',
[5] = '勅捗沈珍朕陳賃鎮追椎墜通痛塚漬坪爪鶴低呈廷弟定底抵邸亭貞帝訂庭逓停偵堤提程艇締諦泥的笛摘滴適敵溺迭哲鉄徹撤天典店点展添転塡田伝殿電斗吐妬徒途都渡塗賭土奴努度怒刀冬灯当投豆東到逃倒凍唐島桃討透党悼盗陶塔搭棟湯痘登答等筒統稲踏糖頭謄藤闘騰同洞胴動堂童道働銅導瞳峠匿特得督徳篤毒独読栃凸突届屯豚頓貪鈍曇丼那奈内梨謎鍋南軟難二尼弐匂肉虹日入乳尿任妊忍認寧熱年念捻粘燃悩納能脳農濃把波派破覇馬婆罵拝杯背肺俳配排敗廃輩売倍梅培陪媒買賠白伯拍泊迫剝舶博薄麦漠縛爆箱箸畑肌八鉢発髪伐抜罰閥反半氾犯帆汎伴判坂阪板版班畔般販斑飯搬煩頒範繁藩晩番蛮盤比皮妃否批彼披肥非卑飛疲秘被悲扉費碑罷避尾眉美備微鼻膝肘匹必泌筆姫百氷表俵票評漂標苗秒病描猫品浜貧賓頻敏瓶不夫父付布扶府怖阜附訃負',
[6] = '赴浮婦符富普腐敷膚賦譜侮武部舞封風伏服副幅復福腹複覆払沸仏物粉紛雰噴墳憤奮分文聞丙平兵併並柄陛閉塀幣弊蔽餅米壁璧癖別蔑片辺返変偏遍編弁辛便勉歩保哺捕補舗母募墓慕暮簿方包芳邦奉宝抱放法泡胞俸倣峰砲崩訪報蜂豊飽褒縫亡乏忙坊妨忘防房肪某冒剖紡望傍帽棒貿貌暴膨謀頰北木朴牧睦僕墨撲没勃堀本奔翻凡盆麻摩磨魔毎妹枚昧埋幕膜枕又末抹万満慢漫未味魅岬密蜜脈妙民眠矛務無夢霧娘名命明迷冥盟銘鳴滅免面綿麺茂模毛妄盲耗猛網目黙門紋問匁冶夜野弥厄役約訳薬躍闇由油喩愉諭輸',
[7] = '癒唯友有勇幽悠郵湧猶裕遊雄誘憂融優与予余誉預幼用羊妖洋要容庸揚揺葉陽溶腰様瘍踊窯養擁謡曜抑沃浴欲翌翼拉裸羅来雷頼絡落酪辣乱卵覧濫藍欄吏利里理痢裏履璃離陸立律慄略柳流留竜粒隆硫侶旅虜慮了両良料涼猟陵量僚領寮療瞭糧力緑林厘倫輪隣臨瑠涙累塁類令礼冷励戻例鈴零霊隷齢麗暦歴列劣烈裂恋連廉練錬呂炉賂路露老労弄郎朗浪廊楼漏籠六録麓論和話賄脇惑枠湾腕',
}
local x = split((frame.args['list'] and frame.args['list'] or kj[tonumber(frame.args[1])]), '')
for i, hz in ipairs(x) do
local content = mw.title.new(hz):getContent() or ''
local jar_ins = match(content, '{{ja%-readings([^}]+)}}')
local jar_new = mw.getCurrentFrame():preprocess('{{#invoke:User:Suzukaze-c/02|test_13' .. jar_ins .. '}}')
local jar_old = '{{ja-readings' .. jar_ins .. '}}'
local ok = (jar_old == jar_new)
mw.log('__' .. jar_old .. '__\n__' .. jar_new .. '__\n\n\n')
local url = export.test_10(content, 'Readings', hz)
if not ok then
table.insert(out, '[' .. url .. ' ' .. hz .. '、]')
end
end
return '<span lang="ja" class="plainlinks">' .. table.concat(out, '') .. '</span>'
end
function export.test_13(frame)
local a = frame.args
local hz = mw.title.getCurrentTitle().text
local yomi = mw.loadData('Module:ja/data/jouyou-yomi').yomi[hz]
local out = {}
local order = {'goon', 'kanon', 'toon', 'kanyoon', 'soon', 'on', 'kun', 'nanori'}
local processed = {}
local missing = {}
for type, text in pairs(a) do
text = replace(text, '%s*%([a-zāīūēō%.%-]%)', '')
text = replace(text, '{{non%-joyo%-reading}}%s*', '')
text = replace(text, '%s*{{q[a-z]*|non%-%[%[w:Jōyō kanji|Jōyō%]%] reading}}', '')
text = replace(text, ',%s*{{q[a-z]*|historical}}', '<')
text = replace(text, '%[%[(' .. hz .. '[ぁ-ー]+)%]%]', '')
text = replace(text, '{{[jal|-]+(' .. hz .. '[ぁ-ー]+)}}', '')
text = replace(text, '%[%[([^%]|]+)|([^%]|]+)%]%]', function(a,b)
return replace(b, '%.', '-')
end
)
text = replace(text, '([^ぁ-ー])%.([^ぁ-ー])', '@')
text = replace(text, '[^ぁ-ー%-<>.]+', '@')
text = replace(text, '(@*)[<>](@*)', '<') -- or something. also, the > is intentional (some entries indeed have backwards arrows)
text = replace(text, '^@', '')
text = replace(text, '@$', '')
text = replace(text, '@', ', ')
if type ~= 'kun' then
text = replace(text, '%-', '')
end
processed[type] = text
end
table.insert(out, '{{ja-readings')
for _, type in ipairs(order) do
if processed[type] then
table.insert(out, '|'..type..'='..processed[type])
end
end
table.insert(out, '}}')
out = table.concat(out, '\n')
if yomi then
for reading, type in pairs(yomi) do
reading = require('Module:string utilities').pattern_escape(reading)
if type == 1 or type == 3 then
reading = require('Module:ja').kata_to_hira(reading)
end
if not (
find(out, '[^ぁ-ー]' .. reading .. '[^ぁ-ー]')
--[[
or
find(out, '=' .. reading .. '[^ぁ-ー]')
or
find(out, '[^ぁ-ー]' .. reading .. '\n')
or
find(out, '=' .. reading .. '\n')
]]
) then
table.insert(missing, reading)
end
end
end
missing = table.concat(missing, ', ') -- hastemplate:ja-readings insource:/\{\{ja-readings[^}]+\}\}[ぁ-ー]/
return out .. replace(missing, '%%', '')
end
function export.test_14(frame)
local text = frame.args[1]
local entries = export.test_18(text)
local x = {}
table.insert(x, '{| class="wikitable sortable"')
table.insert(x, '! - || -')
for entry in entries do
local content = mw.title.new(entry):getContent() or ''
local attention = match(content, '{{attention|([^}]+)}}') or match(content, '{{attn|([^}]+)}}') or '<mark>?</mark>'
local line = '![['..entry..']]\n|'..replace(attention, '|', '¦')
table.insert(x, line)
end
table.insert(x, '|}')
return table.concat(x, '\n|-\n')
end
-- [[mediawiki:Help:Extension:Kartographer]]
function export.make_map(data)
local frame = mw.getCurrentFrame()
local to_json = require("Module:Json").jsonValueFromValue
local mapframe = frame:extensionTag({
name = "mapframe",
content = to_json(data),
args = {
width = 1000,
height = 800,
zoom = 4,
latitude = 30,
longitude = 105,
frameless = true,
align = "center",
}
})
return mapframe
end
function export.map_test_data()
local data = {}
table.insert(data, {
type = "Feature",
geometry = {
type = "Point",
coordinates = { 114, 23 },
},
properties = {
title = "Kwangtung",
fill = "#07c63e",
}
})
return data
end
function export.test_16(frame)
local a1 = split('豆斧魚鋸晶狼箆泌達目劣属主履絛症狭獺香沿用草除鴻章湿膚新神宙浮欽抵参六販酪猿酢西狐艦鷴蜃鶺老緑兼星鹿楔腸鼠匹殉嬢法十拳娘奴卵脱円塾私烏鳥僚漢絆嘘豚刈俺親編抗貴二酔海体以抒友傘合相大總总訣鬪鮫女岩牛刀泉水竜鞭雷羊茸獣心妃霧敵王亀青蠍蛇蝮花鏡嘴口茨冠鎌鎖觜喙鰐森石屍沌翼蕉芭懽歓蝗嵐腕紙狗猪莨烟蜱曳苣萵祥欝蝨虱飯障坂蝶鋲癬鸛燕糞鬆癇癲鼈託鯰徳德商料煙接嚔府戯叔蚣蛺椒胡颪易貿鵆鴟鵂鶹鵩炬鞋藁沓誕曼鰰鰤鳶虫兜澄黽飴哥蟇雛斑樺那支藍奇蹄駒楼嫁徨彷喰郭囃炎蘇蒻重菎蒟酒麦弾陸光組犀過僕氈蕈単婦婚複貝鵡鸚樹台燭辞指臨', '')
local a2 = split('再村烧燒誓尸変縄鵜国公後午爵氷弟家篇餅蝿帳几弁鵲鰌泥哈土艮蛍勝君長集現貘魄刃葉子姫隼妻手敬莓苺肩爺致鯢束約工夢獏打奕語粤朏湾之詰湖瓶死隕蛭産揚寺楽雅素痛頭前度知風邪売卸箱清魂準標鍵無軸枢団教宗術美聖人成鶚身化牝雌酸硫亜亞帽高雉梟客刺義超鋼羽羆狒猴剣廻輪彦車灯提赤明名南北東田陽林川悪空時鯨血地年天闇黒白熊蛙狸鶏小猫犬金木月日男九八七五三一四秋冬徴次魔髄割先氏彼流寓箸夫丈母婆父姉兄農乳然万千百廿仮分至鯱巨沖洞唇鶩半点濁諺薬缶原雨畳誌雑瞳溶狹陿囘回徊黥史樋秩鵐潤由考族傷聡左右彙猬蝟悁捐涓貫慣溝鼷妹強味棒閏鶉鰯似税審会幻喃哪劫永野干瓜蕃舘古醤声柔粥蝋折栓蛾想精椰燃謎密塔饗歌状堅塁蚤旧鴫鷸戌納紐祕躄叭錐顏繪拉費琰未島山沼癸壬辛庚己甲丙丁乙戊卯亥酉申寅辰丑巳虎兎上虜良遍磯秒育黍雁擂鈴杏醋已蛼靨燧梨刻彫炉艫鶴馭幼蛬蠅鮭蓄碇錨釈粒肝恐凰鯧俎社魁雀称染聯隊菱斗灸可眠宅梢柘嚢異椅鎮都裟袈覧拇鵥痘鵄瀝歴紫昇欱規楓顕掃冷尊診淫卓蜂穿館駆文也眉髟豬貍第鯛講蝌訓鮮雹驎騏洋疑吃符嘆茘函鷽弗鏨鑢資捕睦勒隷堕黴痢痳淋争菫疱甚瞬鰹鱓魳魣克簀葭葦瀬吉好渡谷橘篠貞立迎浦桂諫崎臘真鶸鵟鸇鶻鷲鷹挑燈條鳧鴇鴨康鴦鴛昏郎責力張旦面睛螢翆翠翡爆錦鉛沃銀砒銅鉄珪硼窒枘坊裴齣湯銭形罔魎魍莫厨雎龍侍観孫殺聚又鈷槃涅戈戟鋒鉾矛刹祠叢試模肛惚繭摸韆鞦計鍛盗姥瘴儡傀珀琥珊桔穴師薩菩乎愚舵梶牽織袍褞歇浴倶夏春摩獾房俗腱帯派鵝鵞選豹哲橽芃隈余積差和被憲拡卿朕黙薇炭紀識僧方腹駝駱改導綱消羚得鱇鮟瑁瑇玳澹憺啗亶膽耽鄲丹擔戴貸帶詒逮隶瑳磋忖焠淬倅毳萃衰竄趣徂粗從聰朿賜此辭蜯倍堋啜脣淳純捶吹牀撞窻喘椽傳俶觸黜畜杵處芻廚出臭讎愁疇綢稠抽寵崇蟲衝憃舂熾斥翅齒侈哆恥弛馳踟池遲坻螭魑逞騁乘丞承盛呈懲稱瞠晨忱瞋唱悵鬯倡敞裳嘗諂單纏躔蝉廛豺儕察槎插曾測廁册曹操藏倉粲蠶殘慙驂菜蔡彩采簿卜補哺播簸跛薄帛撥番枋並秉稟禀冰表變諞卞辯釆貶邊畢必蹕庇璧蔽臂閉嬖斃幣敝婢辟鄙逼崩奔悖孛杯卑碑陂報瀑抱暴保寶褒胞傍蚌邦辦阪頒拜稗敗霸罷茇奧澳媼熬岸案按鞍艾愛隘埃榜卬端橋轤轆置引普靫犯詞済角震免態毒剖科博振本俄棍柴倫鯆甸勃剌伯賓律匈威抹班瑞坡墺印爾阿陀蘭転濯洗錬副領統軟背反頂両放側擦裂拗便宝磨蝎秤処獅牡蜍蟾駃夬是騠騾驢縞袋蒼鷺鵠紋聞瓦運線幹螟鴿鳩恋量探連宇針荒鵯栗蟀蟋螽斯蝣蜉津蛉蜻鵇禿鳴啼麋双娑訳叉夜羅弥冉耶瓊呂芸薙迦智軻弉伊佐翁利追了最始開欲韓制造創速薔続喚融通式備果能召謀暗壁鬼鉱罠行飛蛛蜘蝓蛞殿民貂兔者忍麟猟馬門漠蟹築射儀秘雪説策類爬昆使座耗落洒駄書中怪繁妖樂市加寿闍毘比餡巴奈独注縦横紅獱帥', '')
local x = {}
for i, hz in ipairs(a1) do
local content = mw.title.new(hz):getContent() or ''
local r = match(content, '|kun=([^|}]+)')
if r then
local rsplit = split(r, ', *')
local rnew = {}
for _, y in ipairs(rsplit) do
if match(y, '[うくぐすずつづぬふぶむゆる]%-$') then
table.insert(rnew, y)
end
end
if #rnew > 0 then
r = '{{ja-readings|kun=' .. table.concat(rnew, ', ') .. '|pagename=' .. hz .. '}}'
table.insert(x, '\n==[[' .. hz .. ']]==\n' .. frame:preprocess(r))
end
else
mw.log(hz)
end
end
return table.concat(x, ' ')
end
function export.test_17()
local data = mw.loadData('Module:ja/data/jouyou-yomi').yomi
local ret = {
'<mark>red link</mark>',
'<del>[[ok]]</del>',
'<u>missing ja-kanjitab alt or ja-see</u>',
'<i>reading not found on kanji-spelling entry</i>',
'<b>ja-see to kanji on on kanji-spelling entry</b>',
}
for z, ys in pairs(data) do
local t = ''
for y, type in pairs(ys) do
if type == 1 or type == 3 then
-- TODO
elseif type == 4 then
-- rare kun'yomi. do not do anything for now.
elseif type == 2 then
local yomiG, okuriG = string.match(y, '^(.+)%-(.*)$')
local o = string.format('%s%s', z, okuriG)
local oRubied = string.format('<ruby>%s<rt>%s</rt></ruby>%s', z, yomiG, okuriG)
local yStripped = string.gsub(y, '%-', '')
local link = string.format('[[%s|%s]]<sub>[[%s]]</sub>', o, oRubied, yStripped)
local oC = mw.title.new(o):getContent()
local yC = mw.title.new(yStripped):getContent()
if not oC or not yC then
link = '<mark>' .. link .. '</mark>'
else
oC = string.gsub(oC, '%{%{ja%-readings[^}]+%}%}', '')
local oFound = string.match(oC, '%{%{ja%-see%|' .. yStripped)
local oFoundSee = string.match(oC, '%{%{ja%-see%|([^}]*)')
local yFound = string.match(yC, '%{%{ja%-kanjitab%|[^}]*' .. o)
local oFoundReading = string.match(string.gsub(oC, '[ .-]', ''), yStripped)
if oFound and yFound then
link = '<del>' .. link .. '</del>'
elseif oFound or yFound then
link = '<u>' .. link .. '</u>'
end
if not oFoundReading then
link = '<i>' .. link .. '</i>'
end
if oFoundSee and match(oFoundSee, '[㐀-鿕]') then
link = link .. '<b><sup>' .. oFoundSee .. '<sup></b>'
end
end
t = t .. link
end
end
table.insert(ret, t)
end
return table.concat(ret) .. require("Module:TemplateStyles")("Template:sandbox/styles.css")
end
function export.test_18(text)
local entries = {}
--[[
text = replace(text, ".+are in this category, out of %d+ total%.%s+", "")
text = replace(text, "\nCategory:.+", "")
for entry in itermatch(text, " +([^\n]+)") do
table.insert(entries, entry)
end
]]
for entry in itermatch(text, ' ([^\n]+)') do
table.insert(entries, entry)
end
return entries
end
function export.test_19(frame)
local text = frame.args[1]
local outZhSee = {}
local outZhForms = {}
local entries = export.test_18(text)
for _, entry in ipairs(entries) do
local content = mw.title.new(entry):getContent() or ''
local zhSees = {}
local zhFormss = {}
for template in itermatch(content, "(%{%{zh%-see[^\n]*%}%})") do
template = mw.getCurrentFrame():preprocess(template)
table.insert(zhSees, template)
end
for template in itermatch(content, "(%{%{zh%-forms[^\n]*%}%})") do
template = replace(template, '%}%}', '|t=' .. entry .. '}}')
template = mw.getCurrentFrame():preprocess(template)
table.insert(zhFormss, template)
end
if #zhSees ~= 0 then
table.insert(outZhSee, "=[https://en.wiktionary.org/w/index.php?title=" .. entry .. "&action=edit " .. entry .. "] [https://en.wiktionary.org/w/index.php?title=" .. entry .. "&action=history hist]=\n")
for _, template in ipairs(zhSees) do
table.insert(outZhSee, template .. "\n")
end
end
if #zhFormss ~= 0 then
table.insert(outZhForms, "=[https://en.wiktionary.org/w/index.php?title=" .. entry .. "&action=edit " .. entry .. "] [https://en.wiktionary.org/w/index.php?title=" .. entry .. "&action=history hist]=\n")
for _, template in ipairs(zhFormss) do
table.insert(outZhForms, template .. "\n")
end
table.insert(outZhForms, "<div style='clear:both'></div>\n")
end
end
return table.concat(outZhSee) .. "\n" .. table.concat(outZhForms)
end
return export