#include "wx/log.h"
#include "wx/intl.h"
#include "wx/encinfo.h"
+ #include "wx/wxcrtvararg.h"
#endif //WX_PRECOMP
#include "wx/msw/private.h"
bool wxNativeEncodingInfo::FromString(const wxString& s)
{
- wxStringTokenizer tokenizer(s, _T(";"));
+ wxStringTokenizer tokenizer(s, wxT(";"));
wxString encid = tokenizer.GetNextToken();
}
else
{
- if ( wxSscanf(tmp, _T("%u"), &charset) != 1 )
+ if ( wxSscanf(tmp, wxT("%u"), &charset) != 1 )
{
// should be a number!
return false;
// we don't have any choice but to use the raw value
<< (long)encoding
#endif // wxUSE_FONTMAP/!wxUSE_FONTMAP
- << _T(';') << facename;
+ << wxT(';') << facename;
// ANSI_CHARSET is assumed anyhow
if ( charset != ANSI_CHARSET )
{
- s << _T(';') << charset;
+ s << wxT(';') << charset;
}
return s;
bool wxGetNativeFontEncoding(wxFontEncoding encoding,
wxNativeEncodingInfo *info)
{
- wxCHECK_MSG( info, false, _T("bad pointer in wxGetNativeFontEncoding") );
+ wxCHECK_MSG( info, false, wxT("bad pointer in wxGetNativeFontEncoding") );
if ( encoding == wxFONTENCODING_DEFAULT )
{
wxZeroMemory(lf); // all default values
lf.lfCharSet = (BYTE)info.charset;
- wxStrncpy(lf.lfFaceName, info.facename, WXSIZEOF(lf.lfFaceName));
+ wxStrlcpy(lf.lfFaceName, info.facename.c_str(), WXSIZEOF(lf.lfFaceName));
HFONT hfont = ::CreateFontIndirect(&lf);
if ( !hfont )
switch ( cs )
{
default:
- wxFAIL_MSG( _T("unexpected Win32 charset") );
+ wxFAIL_MSG( wxT("unexpected Win32 charset") );
// fall through and assume the system charset
case DEFAULT_CHARSET: